Privacy Preservation in Federated Learning: An insightful survey from
the GDPR Perspective
- URL: http://arxiv.org/abs/2011.05411v5
- Date: Thu, 18 Mar 2021 12:32:28 GMT
- Title: Privacy Preservation in Federated Learning: An insightful survey from
the GDPR Perspective
- Authors: Nguyen Truong, Kai Sun, Siyao Wang, Florian Guitton, Yike Guo
- Abstract summary: Article is dedicated to surveying on the state-of-the-art privacy techniques, which can be employed in Federated learning.
Recent research has demonstrated that retaining data and on computation in FL is not enough for privacy-guarantee.
This is because ML model parameters exchanged between parties in an FL system, which can be exploited in some privacy attacks.
- Score: 10.901568085406753
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Along with the blooming of AI and Machine Learning-based applications and
services, data privacy and security have become a critical challenge.
Conventionally, data is collected and aggregated in a data centre on which
machine learning models are trained. This centralised approach has induced
severe privacy risks to personal data leakage, misuse, and abuse. Furthermore,
in the era of the Internet of Things and big data in which data is essentially
distributed, transferring a vast amount of data to a data centre for processing
seems to be a cumbersome solution. This is not only because of the difficulties
in transferring and sharing data across data sources but also the challenges on
complying with rigorous data protection regulations and complicated
administrative procedures such as the EU General Data Protection Regulation
(GDPR). In this respect, Federated learning (FL) emerges as a prospective
solution that facilitates distributed collaborative learning without disclosing
original training data whilst naturally complying with the GDPR. Recent
research has demonstrated that retaining data and computation on-device in FL
is not sufficient enough for privacy-guarantee. This is because ML model
parameters exchanged between parties in an FL system still conceal sensitive
information, which can be exploited in some privacy attacks. Therefore, FL
systems shall be empowered by efficient privacy-preserving techniques to comply
with the GDPR. This article is dedicated to surveying on the state-of-the-art
privacy-preserving techniques which can be employed in FL in a systematic
fashion, as well as how these techniques mitigate data security and privacy
risks. Furthermore, we provide insights into the challenges along with
prospective approaches following the GDPR regulatory guidelines that an FL
system shall implement to comply with the GDPR.
Related papers
- Privacy in Federated Learning [0.0]
Federated Learning (FL) represents a significant advancement in distributed machine learning.
This chapter delves into the core privacy concerns within FL, including the risks of data reconstruction, model inversion attacks, and membership inference.
It examines the trade-offs between model accuracy and privacy, emphasizing the importance of balancing these factors in practical implementations.
arXiv Detail & Related papers (2024-08-12T18:41:58Z) - Lancelot: Towards Efficient and Privacy-Preserving Byzantine-Robust Federated Learning within Fully Homomorphic Encryption [10.685816010576918]
We propose Lancelot, an innovative and computationally efficient BRFL framework that employs fully homomorphic encryption (FHE) to safeguard against malicious client activities while preserving data privacy.
Our extensive testing, which includes medical imaging diagnostics and widely-used public image datasets, demonstrates that Lancelot significantly outperforms existing methods, offering more than a twenty-fold increase in processing speed, all while maintaining data privacy.
arXiv Detail & Related papers (2024-08-12T14:48:25Z) - Privacy-Enhancing Collaborative Information Sharing through Federated
Learning -- A Case of the Insurance Industry [1.8092553911119764]
The report demonstrates the benefits of harnessing the value of Federated Learning (FL) to learn a single model across multiple insurance industry datasets.
FL addresses two of the most pressing concerns: limited data volume and data variety, which are caused by privacy concerns.
During each round of FL, collaborators compute improvements on the model using their local private data, and these insights are combined to update a global model.
arXiv Detail & Related papers (2024-02-22T21:46:24Z) - Federated Learning Empowered by Generative Content [55.576885852501775]
Federated learning (FL) enables leveraging distributed private data for model training in a privacy-preserving way.
We propose a novel FL framework termed FedGC, designed to mitigate data heterogeneity issues by diversifying private data with generative content.
We conduct a systematic empirical study on FedGC, covering diverse baselines, datasets, scenarios, and modalities.
arXiv Detail & Related papers (2023-12-10T07:38:56Z) - PS-FedGAN: An Efficient Federated Learning Framework Based on Partially
Shared Generative Adversarial Networks For Data Privacy [56.347786940414935]
Federated Learning (FL) has emerged as an effective learning paradigm for distributed computation.
This work proposes a novel FL framework that requires only partial GAN model sharing.
Named as PS-FedGAN, this new framework enhances the GAN releasing and training mechanism to address heterogeneous data distributions.
arXiv Detail & Related papers (2023-05-19T05:39:40Z) - Federated Learning with Privacy-Preserving Ensemble Attention
Distillation [63.39442596910485]
Federated Learning (FL) is a machine learning paradigm where many local nodes collaboratively train a central model while keeping the training data decentralized.
We propose a privacy-preserving FL framework leveraging unlabeled public data for one-way offline knowledge distillation.
Our technique uses decentralized and heterogeneous local data like existing FL approaches, but more importantly, it significantly reduces the risk of privacy leakage.
arXiv Detail & Related papers (2022-10-16T06:44:46Z) - Do Gradient Inversion Attacks Make Federated Learning Unsafe? [70.0231254112197]
Federated learning (FL) allows the collaborative training of AI models without needing to share raw data.
Recent works on the inversion of deep neural networks from model gradients raised concerns about the security of FL in preventing the leakage of training data.
In this work, we show that these attacks presented in the literature are impractical in real FL use-cases and provide a new baseline attack.
arXiv Detail & Related papers (2022-02-14T18:33:12Z) - Distributed Machine Learning and the Semblance of Trust [66.1227776348216]
Federated Learning (FL) allows the data owner to maintain data governance and perform model training locally without having to share their data.
FL and related techniques are often described as privacy-preserving.
We explain why this term is not appropriate and outline the risks associated with over-reliance on protocols that were not designed with formal definitions of privacy in mind.
arXiv Detail & Related papers (2021-12-21T08:44:05Z) - A Principled Approach to Data Valuation for Federated Learning [73.19984041333599]
Federated learning (FL) is a popular technique to train machine learning (ML) models on decentralized data sources.
The Shapley value (SV) defines a unique payoff scheme that satisfies many desiderata for a data value notion.
This paper proposes a variant of the SV amenable to FL, which we call the federated Shapley value.
arXiv Detail & Related papers (2020-09-14T04:37:54Z) - Federated Extra-Trees with Privacy Preserving [20.564530457026976]
We propose a novel privacy-preserving machine learning model named Federated Extra-Trees.
A secure multi-institutional machine learning system was developed to provide superior performance.
arXiv Detail & Related papers (2020-02-18T01:15:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.