Forgetting Any Data at Any Time: A Theoretically Certified Unlearning Framework for Vertical Federated Learning
- URL: http://arxiv.org/abs/2502.17081v1
- Date: Mon, 24 Feb 2025 11:52:35 GMT
- Title: Forgetting Any Data at Any Time: A Theoretically Certified Unlearning Framework for Vertical Federated Learning
- Authors: Linian Wang, Leye Wang,
- Abstract summary: We introduce the first VFL framework with theoretically guaranteed unlearning capabilities.<n>Unlike prior approaches, our solution is model- and data-agnostic, offering universal compatibility.<n>Our framework supports asynchronous unlearning, eliminating the need for all parties to be simultaneously online during the forgetting process.
- Score: 8.127710748771992
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Privacy concerns in machine learning are heightened by regulations such as the GDPR, which enforces the "right to be forgotten" (RTBF), driving the emergence of machine unlearning as a critical research field. Vertical Federated Learning (VFL) enables collaborative model training by aggregating a sample's features across distributed parties while preserving data privacy at each source. This paradigm has seen widespread adoption in healthcare, finance, and other privacy-sensitive domains. However, existing VFL systems lack robust mechanisms to comply with RTBF requirements, as unlearning methodologies for VFL remain underexplored. In this work, we introduce the first VFL framework with theoretically guaranteed unlearning capabilities, enabling the removal of any data at any time. Unlike prior approaches -- which impose restrictive assumptions on model architectures or data types for removal -- our solution is model- and data-agnostic, offering universal compatibility. Moreover, our framework supports asynchronous unlearning, eliminating the need for all parties to be simultaneously online during the forgetting process. These advancements address critical gaps in current VFL systems, ensuring compliance with RTBF while maintaining operational flexibility.We make all our implementations publicly available at https://github.com/wangln19/vertical-federated-unlearning.
Related papers
- Federated Unlearning Made Practical: Seamless Integration via Negated Pseudo-Gradients [3.12131298354022]
This paper introduces a novel method that leverages negated Pseudo-gradients Updates for Federated Unlearning (PUF)
Our approach only uses standard client model updates, anyway employed during regular FL rounds, and interprets them as pseudo-gradients.
Unlike state-of-the-art mechanisms, PUF seamlessly integrates with FL, incurs no additional computational and communication overhead beyond standard FL rounds, and supports concurrent unlearning requests.
arXiv Detail & Related papers (2025-04-08T09:05:33Z) - Vertical Federated Learning in Practice: The Good, the Bad, and the Ugly [42.31182713177944]
This survey analyzes the real-world data distributions in potential Vertical Federated Learning (VFL) applications.<n>We propose a novel data-oriented taxonomy of VFL algorithms based on real VFL data distributions.<n>Based on these observations, we outline key research directions aimed at bridging the gap between current VFL research and real-world applications.
arXiv Detail & Related papers (2025-02-12T07:03:32Z) - Advances in APPFL: A Comprehensive and Extensible Federated Learning Framework [1.4206132527980742]
Federated learning (FL) is a distributed machine learning paradigm enabling collaborative model training while preserving data privacy.
We present the recent advances in developing APPFL, a framework and benchmarking suite for federated learning.
We demonstrate the capabilities of APPFL through extensive experiments evaluating various aspects of FL, including communication efficiency, privacy preservation, computational performance, and resource utilization.
arXiv Detail & Related papers (2024-09-17T22:20:26Z) - FedQUIT: On-Device Federated Unlearning via a Quasi-Competent Virtual Teacher [4.291269657919828]
Federated Learning (FL) promises better privacy guarantees for individuals' data when machine learning models are collaboratively trained.
When an FL participant exercises its right to be forgotten, i.e., to detach from the FL framework it has participated, the FL solution should perform all the necessary steps.
We propose FedQUIT, a novel algorithm that uses knowledge distillation to scrub the contribution of the forgetting data from an FL global model.
arXiv Detail & Related papers (2024-08-14T14:36:28Z) - UIFV: Data Reconstruction Attack in Vertical Federated Learning [5.404398887781436]
Vertical Federated Learning (VFL) facilitates collaborative machine learning without the need for participants to share raw private data.<n>Recent studies have revealed privacy risks where adversaries might reconstruct sensitive features through data leakage during the learning process.<n>Our work exposes severe privacy vulnerabilities within VFL systems that pose real threats to practical VFL applications.
arXiv Detail & Related papers (2024-06-18T13:18:52Z) - A Comprehensive Study on Model Initialization Techniques Ensuring
Efficient Federated Learning [0.0]
Federated learning(FL) has emerged as a promising paradigm for training machine learning models in a distributed and privacy-preserving manner.
The choice of methods used for models plays a crucial role in the performance, convergence speed, communication efficiency, privacy guarantees of federated learning systems.
Our research meticulously compares, categorizes, and delineates the merits and demerits of each technique, examining their applicability across diverse FL scenarios.
arXiv Detail & Related papers (2023-10-31T23:26:58Z) - Heterogeneous Decentralized Machine Unlearning with Seed Model
Distillation [47.42071293545731]
Information security legislation endowed users with unconditional rights to be forgotten by trained machine learning models.
We design a decentralized unlearning framework called HDUS, which uses distilled seed models to construct erasable ensembles for all clients.
arXiv Detail & Related papers (2023-08-25T09:42:54Z) - Federated Learning and Meta Learning: Approaches, Applications, and
Directions [94.68423258028285]
In this tutorial, we present a comprehensive review of FL, meta learning, and federated meta learning (FedMeta)
Unlike other tutorial papers, our objective is to explore how FL, meta learning, and FedMeta methodologies can be designed, optimized, and evolved, and their applications over wireless networks.
arXiv Detail & Related papers (2022-10-24T10:59:29Z) - Vertical Semi-Federated Learning for Efficient Online Advertising [50.18284051956359]
Semi-VFL (Vertical Semi-Federated Learning) is proposed to achieve a practical industry application fashion for VFL.
We build an inference-efficient single-party student model applicable to the whole sample space.
New representation distillation methods are designed to extract cross-party feature correlations for both the overlapped and non-overlapped data.
arXiv Detail & Related papers (2022-09-30T17:59:27Z) - Secure Bilevel Asynchronous Vertical Federated Learning with Backward
Updating [159.48259714642447]
Vertical scalable learning (VFL) attracts increasing attention due to the demands of multi-party collaborative modeling and concerns of privacy leakage.
We propose a novel bftextlevel parallel architecture (VF$bfB2$), under which three new algorithms, including VF$B2$, are proposed.
arXiv Detail & Related papers (2021-03-01T12:34:53Z) - Edge-assisted Democratized Learning Towards Federated Analytics [67.44078999945722]
We show the hierarchical learning structure of the proposed edge-assisted democratized learning mechanism, namely Edge-DemLearn.
We also validate Edge-DemLearn as a flexible model training mechanism to build a distributed control and aggregation methodology in regions.
arXiv Detail & Related papers (2020-12-01T11:46:03Z) - A Principled Approach to Data Valuation for Federated Learning [73.19984041333599]
Federated learning (FL) is a popular technique to train machine learning (ML) models on decentralized data sources.
The Shapley value (SV) defines a unique payoff scheme that satisfies many desiderata for a data value notion.
This paper proposes a variant of the SV amenable to FL, which we call the federated Shapley value.
arXiv Detail & Related papers (2020-09-14T04:37:54Z) - Wireless Communications for Collaborative Federated Learning [160.82696473996566]
Internet of Things (IoT) devices may not be able to transmit their collected data to a central controller for training machine learning models.
Google's seminal FL algorithm requires all devices to be directly connected with a central controller.
This paper introduces a novel FL framework, called collaborative FL (CFL), which enables edge devices to implement FL with less reliance on a central controller.
arXiv Detail & Related papers (2020-06-03T20:00:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.