A New Dimensionality Reduction Method Based on Hensel's Compression for
Privacy Protection in Federated Learning
- URL: http://arxiv.org/abs/2205.02089v1
- Date: Sun, 1 May 2022 23:52:16 GMT
- Title: A New Dimensionality Reduction Method Based on Hensel's Compression for
Privacy Protection in Federated Learning
- Authors: Ahmed El Ouadrhiri, Ahmed Abdelhadi
- Abstract summary: We propose two layers of privacy protection approach to overcome the limitations of existing DP-based approaches.
The first layer reduces the dimension of the training dataset based on Hensel's Lemma.
The second layer applies DP to the compressed dataset generated by the first layer.
- Score: 1.0152838128195467
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Differential privacy (DP) is considered a de-facto standard for protecting
users' privacy in data analysis, machine, and deep learning. Existing DP-based
privacy-preserving training approaches consist of adding noise to the clients'
gradients before sharing them with the server. However, implementing DP on the
gradient is not efficient as the privacy leakage increases by increasing the
synchronization training epochs due to the composition theorem. Recently
researchers were able to recover images used in the training dataset using
Generative Regression Neural Network (GRNN) even when the gradient was
protected by DP. In this paper, we propose two layers of privacy protection
approach to overcome the limitations of the existing DP-based approaches. The
first layer reduces the dimension of the training dataset based on Hensel's
Lemma. We are the first to use Hensel's Lemma for reducing the dimension (i.e.,
compress) of a dataset. The new dimensionality reduction method allows reducing
the dimension of a dataset without losing information since Hensel's Lemma
guarantees uniqueness. The second layer applies DP to the compressed dataset
generated by the first layer. The proposed approach overcomes the problem of
privacy leakage due to composition by applying DP only once before the
training; clients train their local model on the privacy-preserving dataset
generated by the second layer. Experimental results show that the proposed
approach ensures strong privacy protection while achieving good accuracy. The
new dimensionality reduction method achieves an accuracy of 97%, with only 25 %
of the original data size.
Related papers
- Pre-training Differentially Private Models with Limited Public Data [54.943023722114134]
differential privacy (DP) is a prominent method to gauge the degree of security provided to the models.
DP is yet not capable of protecting a substantial portion of the data used during the initial pre-training stage.
We develop a novel DP continual pre-training strategy using only 10% of public data.
Our strategy can achieve DP accuracy of 41.5% on ImageNet-21k, as well as non-DP accuracy of 55.7% and and 60.0% on downstream tasks Places365 and iNaturalist-2021.
arXiv Detail & Related papers (2024-02-28T23:26:27Z) - Conciliating Privacy and Utility in Data Releases via Individual Differential Privacy and Microaggregation [4.287502453001108]
$epsilon$-Differential privacy (DP) is a well-known privacy model that offers strong privacy guarantees.
We propose $epsilon$-individual differential privacy (iDP), which causes less data distortion while providing the same protection as DP to subjects.
We report on experiments that show how our approach can provide strong privacy (small $epsilon$) while yielding protected data that do not significantly degrade the accuracy of secondary data analysis.
arXiv Detail & Related papers (2023-12-21T10:23:18Z) - Differentially Private SGD Without Clipping Bias: An Error-Feedback Approach [62.000948039914135]
Using Differentially Private Gradient Descent with Gradient Clipping (DPSGD-GC) to ensure Differential Privacy (DP) comes at the cost of model performance degradation.
We propose a new error-feedback (EF) DP algorithm as an alternative to DPSGD-GC.
We establish an algorithm-specific DP analysis for our proposed algorithm, providing privacy guarantees based on R'enyi DP.
arXiv Detail & Related papers (2023-11-24T17:56:44Z) - Sparsity-Preserving Differentially Private Training of Large Embedding
Models [67.29926605156788]
DP-SGD is a training algorithm that combines differential privacy with gradient descent.
Applying DP-SGD naively to embedding models can destroy gradient sparsity, leading to reduced training efficiency.
We present two new algorithms, DP-FEST and DP-AdaFEST, that preserve gradient sparsity during private training of large embedding models.
arXiv Detail & Related papers (2023-11-14T17:59:51Z) - GIFD: A Generative Gradient Inversion Method with Feature Domain
Optimization [52.55628139825667]
Federated Learning (FL) has emerged as a promising distributed machine learning framework to preserve clients' privacy.
Recent studies find that an attacker can invert the shared gradients and recover sensitive data against an FL system by leveraging pre-trained generative adversarial networks (GAN) as prior knowledge.
We propose textbfGradient textbfInversion over textbfFeature textbfDomains (GIFD), which disassembles the GAN model and searches the feature domains of the intermediate layers.
arXiv Detail & Related papers (2023-08-09T04:34:21Z) - Differential Privacy Meets Neural Network Pruning [10.77469946354744]
We study the interplay between neural network pruning and differential privacy, through the two modes of parameter updates.
Our experimental results demonstrate how decreasing the parameter space improves differentially private training.
By studying two popular forms of pruning which do not rely on gradients and do not incur an additional privacy loss, we show that random selection performs on par with magnitude-based selection.
arXiv Detail & Related papers (2023-03-08T14:27:35Z) - DataLens: Scalable Privacy Preserving Training via Gradient Compression
and Aggregation [15.63770709526671]
We propose a scalable privacy-preserving generative model DATALENS.
We show that, DATALENS significantly outperforms other baseline DP generative models.
We adapt the proposed TOPAGG approach, which is one of the key building blocks in DATALENS, to DP SGD training.
arXiv Detail & Related papers (2021-03-20T06:14:19Z) - Do Not Let Privacy Overbill Utility: Gradient Embedding Perturbation for
Private Learning [74.73901662374921]
A differentially private model degrades the utility drastically when the model comprises a large number of trainable parameters.
We propose an algorithm emphGradient Embedding Perturbation (GEP) towards training differentially private deep models with decent accuracy.
arXiv Detail & Related papers (2021-02-25T04:29:58Z) - Federated Learning with Sparsification-Amplified Privacy and Adaptive
Optimization [27.243322019117144]
Federated learning (FL) enables distributed agents to collaboratively learn a centralized model without sharing their raw data with each other.
We propose a new FL framework with sparsification-amplified privacy.
Our approach integrates random sparsification with gradient perturbation on each agent to amplify privacy guarantee.
arXiv Detail & Related papers (2020-08-01T20:22:57Z) - Differentially Private Federated Learning with Laplacian Smoothing [72.85272874099644]
Federated learning aims to protect data privacy by collaboratively learning a model without sharing private data among users.
An adversary may still be able to infer the private training data by attacking the released model.
Differential privacy provides a statistical protection against such attacks at the price of significantly degrading the accuracy or utility of the trained models.
arXiv Detail & Related papers (2020-05-01T04:28:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.