Split Without a Leak: Reducing Privacy Leakage in Split Learning
- URL: http://arxiv.org/abs/2308.15783v1
- Date: Wed, 30 Aug 2023 06:28:42 GMT
- Title: Split Without a Leak: Reducing Privacy Leakage in Split Learning
- Authors: Khoa Nguyen, Tanveer Khan and Antonis Michalas
- Abstract summary: We propose a hybrid approach using Split Learning (SL) and Homomorphic Encryption (HE)
On the MIT-BIH dataset, our proposed hybrid approach using SL and HE yields faster training time (about 6 times) and significantly reduced communication overhead (almost 160 times) compared to other HE-based approaches.
- Score: 3.2066885499201176
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The popularity of Deep Learning (DL) makes the privacy of sensitive data more
imperative than ever. As a result, various privacy-preserving techniques have
been implemented to preserve user data privacy in DL. Among various
privacy-preserving techniques, collaborative learning techniques, such as Split
Learning (SL) have been utilized to accelerate the learning and prediction
process. Initially, SL was considered a promising approach to data privacy.
However, subsequent research has demonstrated that SL is susceptible to many
types of attacks and, therefore, it cannot serve as a privacy-preserving
technique. Meanwhile, countermeasures using a combination of SL and encryption
have also been introduced to achieve privacy-preserving deep learning. In this
work, we propose a hybrid approach using SL and Homomorphic Encryption (HE).
The idea behind it is that the client encrypts the activation map (the output
of the split layer between the client and the server) before sending it to the
server. Hence, during both forward and backward propagation, the server cannot
reconstruct the client's input data from the intermediate activation map. This
improvement is important as it reduces privacy leakage compared to other
SL-based works, where the server can gain valuable information about the
client's input. In addition, on the MIT-BIH dataset, our proposed hybrid
approach using SL and HE yields faster training time (about 6 times) and
significantly reduced communication overhead (almost 160 times) compared to
other HE-based approaches, thereby offering improved privacy protection for
sensitive data in DL.
Related papers
- KnowledgeSG: Privacy-Preserving Synthetic Text Generation with Knowledge Distillation from Server [48.04903443425111]
Large language models (LLMs) facilitate many parties to fine-tune LLMs on their own private data.
Existing solutions, such as utilizing synthetic data for substitution, struggle to simultaneously improve performance and preserve privacy.
We propose KnowledgeSG, a novel client-server framework which enhances synthetic data quality and improves model performance while ensuring privacy.
arXiv Detail & Related papers (2024-10-08T06:42:28Z) - Make Split, not Hijack: Preventing Feature-Space Hijacking Attacks in Split Learning [1.6822770693792823]
We introduce a hybrid approach combining Split Learning (SL) and Function Secret Sharing (FSS) to ensure client data privacy.
Our protocols yield promising results, reducing communication overhead by over 2x and training time by over 7x compared to the same model with FSS, without any SL.
arXiv Detail & Related papers (2024-04-14T14:14:31Z) - Enhancing Accuracy-Privacy Trade-off in Differentially Private Split Learning [2.2676798389997863]
Split learning (SL) aims to protect user data privacy by distributing deep models between client-server and keeping private data locally.
Recently proposed model inversion attacks can recover the original data from the smashed data.
A strategy is to adopt differential privacy (DP), which involves safeguarding the smashed data at the expense of some accuracy loss.
arXiv Detail & Related papers (2023-10-22T22:45:13Z) - Love or Hate? Share or Split? Privacy-Preserving Training Using Split
Learning and Homomorphic Encryption [47.86010265348072]
Split learning (SL) is a new collaborative learning technique that allows participants to train machine learning models without the client sharing raw data.
Previous works demonstrated that reconstructing activation maps could result in privacy leakage of client data.
In this paper, we improve upon previous works by constructing a protocol based on U-shaped SL that can operate on homomorphically encrypted data.
arXiv Detail & Related papers (2023-09-19T10:56:08Z) - A More Secure Split: Enhancing the Security of Privacy-Preserving Split Learning [2.853180143237022]
Split learning (SL) is a new collaborative learning technique that allows participants to train machine learning models without the client sharing raw data.
Previous works demonstrated that reconstructing Activation Maps (AMs) could result in privacy leakage of client data.
In this paper, we improve upon previous works by constructing a protocol based on U-shaped SL that can operate on homomorphically encrypted data.
arXiv Detail & Related papers (2023-09-15T18:39:30Z) - Split Ways: Privacy-Preserving Training of Encrypted Data Using Split
Learning [6.916134299626706]
Split Learning (SL) is a new collaborative learning technique that allows participants to train machine learning models without the client sharing raw data.
Previous works demonstrated that reconstructing activation maps could result in privacy leakage of client data.
In this paper, we improve upon previous works by constructing a protocol based on U-shaped SL that can operate on homomorphically encrypted data.
arXiv Detail & Related papers (2023-01-20T19:26:51Z) - Scalable Collaborative Learning via Representation Sharing [53.047460465980144]
Federated learning (FL) and Split Learning (SL) are two frameworks that enable collaborative learning while keeping the data private (on device)
In FL, each data holder trains a model locally and releases it to a central server for aggregation.
In SL, the clients must release individual cut-layer activations (smashed data) to the server and wait for its response (during both inference and back propagation).
In this work, we present a novel approach for privacy-preserving machine learning, where the clients collaborate via online knowledge distillation using a contrastive loss.
arXiv Detail & Related papers (2022-11-20T10:49:22Z) - Server-Side Local Gradient Averaging and Learning Rate Acceleration for
Scalable Split Learning [82.06357027523262]
Federated learning (FL) and split learning (SL) are two spearheads possessing their pros and cons, and are suited for many user clients and large models.
In this work, we first identify the fundamental bottlenecks of SL, and thereby propose a scalable SL framework, coined SGLR.
arXiv Detail & Related papers (2021-12-11T08:33:25Z) - Understanding Clipping for Federated Learning: Convergence and
Client-Level Differential Privacy [67.4471689755097]
This paper empirically demonstrates that the clipped FedAvg can perform surprisingly well even with substantial data heterogeneity.
We provide the convergence analysis of a differential private (DP) FedAvg algorithm and highlight the relationship between clipping bias and the distribution of the clients' updates.
arXiv Detail & Related papers (2021-06-25T14:47:19Z) - TIPRDC: Task-Independent Privacy-Respecting Data Crowdsourcing Framework
for Deep Learning with Anonymized Intermediate Representations [49.20701800683092]
We present TIPRDC, a task-independent privacy-respecting data crowdsourcing framework with anonymized intermediate representation.
The goal of this framework is to learn a feature extractor that can hide the privacy information from the intermediate representations; while maximally retaining the original information embedded in the raw data for the data collector to accomplish unknown learning tasks.
arXiv Detail & Related papers (2020-05-23T06:21:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.