Machine Unlearning of Federated Clusters
- URL: http://arxiv.org/abs/2210.16424v2
- Date: Sat, 1 Jul 2023 03:53:30 GMT
- Title: Machine Unlearning of Federated Clusters
- Authors: Chao Pan, Jin Sima, Saurav Prakash, Vishal Rana, Olgica Milenkovic
- Abstract summary: Federated clustering is an unsupervised learning problem that arises in a number of practical applications, including personalized recommender and healthcare systems.
We introduce, for the first time, the problem of machine unlearning for FC.
We propose an efficient unlearning mechanism for a customized secure FC framework.
- Score: 36.663892269484506
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Federated clustering (FC) is an unsupervised learning problem that arises in
a number of practical applications, including personalized recommender and
healthcare systems. With the adoption of recent laws ensuring the "right to be
forgotten", the problem of machine unlearning for FC methods has become of
significant importance. We introduce, for the first time, the problem of
machine unlearning for FC, and propose an efficient unlearning mechanism for a
customized secure FC framework. Our FC framework utilizes special
initialization procedures that we show are well-suited for unlearning. To
protect client data privacy, we develop the secure compressed multiset
aggregation (SCMA) framework that addresses sparse secure federated learning
(FL) problems encountered during clustering as well as more general problems.
To simultaneously facilitate low communication complexity and secret sharing
protocols, we integrate Reed-Solomon encoding with special evaluation points
into our SCMA pipeline, and prove that the client communication cost is
logarithmic in the vector dimension. Additionally, to demonstrate the benefits
of our unlearning mechanism over complete retraining, we provide a theoretical
analysis for the unlearning performance of our approach. Simulation results
show that the new FC framework exhibits superior clustering performance
compared to previously reported FC baselines when the cluster sizes are highly
imbalanced. Compared to completely retraining K-means++ locally and globally
for each removal request, our unlearning procedure offers an average speed-up
of roughly 84x across seven datasets. Our implementation for the proposed
method is available at https://github.com/thupchnsky/mufc.
Related papers
- On Homomorphic Encryption Based Strategies for Class Imbalance in Federated Learning [4.322339935902437]
Class imbalance in training datasets can lead to bias and poor generalization in machine learning models.
We propose FLICKER, a privacy preserving framework to address issues related to global class imbalance in federated learning.
arXiv Detail & Related papers (2024-10-28T16:35:40Z) - Face Clustering via Early Stopping and Edge Recall [6.847295047818357]
We propose a novel unsupervised face clustering algorithm FC-ES and a novel supervised face clustering algorithm FC-ESER.
Our proposed FC-ES and FC-ESER significantly outperform previous state-of-the-art methods.
arXiv Detail & Related papers (2024-08-24T01:53:02Z) - Scalable Federated Unlearning via Isolated and Coded Sharding [76.12847512410767]
Federated unlearning has emerged as a promising paradigm to erase the client-level data effect.
This paper proposes a scalable federated unlearning framework based on isolated sharding and coded computing.
arXiv Detail & Related papers (2024-01-29T08:41:45Z) - Blockchain-enabled Trustworthy Federated Unlearning [50.01101423318312]
Federated unlearning is a promising paradigm for protecting the data ownership of distributed clients.
Existing works require central servers to retain the historical model parameters from distributed clients.
This paper proposes a new blockchain-enabled trustworthy federated unlearning framework.
arXiv Detail & Related papers (2024-01-29T07:04:48Z) - TurboSVM-FL: Boosting Federated Learning through SVM Aggregation for
Lazy Clients [44.44776028287441]
TurboSVM-FL is a novel federated aggregation strategy that poses no additional computation burden on the client side.
We evaluate TurboSVM-FL on multiple datasets including FEMNIST, CelebA, and Shakespeare.
arXiv Detail & Related papers (2024-01-22T14:59:11Z) - Client Orchestration and Cost-Efficient Joint Optimization for
NOMA-Enabled Hierarchical Federated Learning [55.49099125128281]
We propose a non-orthogonal multiple access (NOMA) enabled HFL system under semi-synchronous cloud model aggregation.
We show that the proposed scheme outperforms the considered benchmarks regarding HFL performance improvement and total cost reduction.
arXiv Detail & Related papers (2023-11-03T13:34:44Z) - When Computing Power Network Meets Distributed Machine Learning: An
Efficient Federated Split Learning Framework [6.871107511111629]
CPN-FedSL is a Federated Split Learning (FedSL) framework over Computing Power Network (CPN)
We build a dedicated model to capture the basic settings and learning characteristics (e.g., latency, flow, convergence)
arXiv Detail & Related papers (2023-05-22T12:36:52Z) - Cali3F: Calibrated Fast Fair Federated Recommendation System [25.388324221293203]
We propose a personalized federated recommendation system training algorithm to improve recommendation performance fairness.
We then adopt a clustering-based aggregation method to accelerate the training process.
Cali3F is a calibrated fast and fair federated recommendation framework.
arXiv Detail & Related papers (2022-05-26T03:05:26Z) - Local Learning Matters: Rethinking Data Heterogeneity in Federated
Learning [61.488646649045215]
Federated learning (FL) is a promising strategy for performing privacy-preserving, distributed learning with a network of clients (i.e., edge devices)
arXiv Detail & Related papers (2021-11-28T19:03:39Z) - Meta Clustering Learning for Large-scale Unsupervised Person
Re-identification [124.54749810371986]
We propose a "small data for big task" paradigm dubbed Meta Clustering Learning (MCL)
MCL only pseudo-labels a subset of the entire unlabeled data via clustering to save computing for the first-phase training.
Our method significantly saves computational cost while achieving a comparable or even better performance compared to prior works.
arXiv Detail & Related papers (2021-11-19T04:10:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.