Secure Federated Clustering
- URL: http://arxiv.org/abs/2205.15564v1
- Date: Tue, 31 May 2022 06:47:18 GMT
- Title: Secure Federated Clustering
- Authors: Songze Li, Sizai Hou, Baturalp Buyukates, Salman Avestimehr
- Abstract summary: SecFC is a secure federated clustering algorithm that simultaneously achieves universal performance.
Each client's private data and the cluster centers are not leaked to other clients and the server.
- Score: 18.37669220755388
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider a foundational unsupervised learning task of $k$-means data
clustering, in a federated learning (FL) setting consisting of a central server
and many distributed clients. We develop SecFC, which is a secure federated
clustering algorithm that simultaneously achieves 1) universal performance: no
performance loss compared with clustering over centralized data, regardless of
data distribution across clients; 2) data privacy: each client's private data
and the cluster centers are not leaked to other clients and the server. In
SecFC, the clients perform Lagrange encoding on their local data and share the
coded data in an information-theoretically private manner; then leveraging the
algebraic structure of the coding, the FL network exactly executes the Lloyd's
$k$-means heuristic over the coded data to obtain the final clustering.
Experiment results on synthetic and real datasets demonstrate the universally
superior performance of SecFC for different data distributions across clients,
and its computational practicality for various combinations of system
parameters. Finally, we propose an extension of SecFC to further provide
membership privacy for all data points.
Related papers
- Federated Clustering: An Unsupervised Cluster-Wise Training for Decentralized Data Distributions [1.6385815610837167]
Federated Cluster-Wise Refinement (FedCRef) involves clients that collaboratively train models on clusters with similar data distributions.
In these groups, clients collaboratively train a shared model representing each data distribution, while continuously refining their local clusters to enhance data association accuracy.
This iterative process allows our system to identify all potential data distributions across the network and develop robust representation models for each.
arXiv Detail & Related papers (2024-08-20T09:05:44Z) - Fair Federated Data Clustering through Personalization: Bridging the Gap between Diverse Data Distributions [2.7905216619150344]
We introduce the idea of personalization in federated clustering. The goal is achieve balance between achieving lower clustering cost and at same time achieving uniform cost across clients.
We propose p-FClus that addresses these goal in a single round of communication between server and clients.
arXiv Detail & Related papers (2024-07-05T07:10:26Z) - Dynamically Weighted Federated k-Means [0.0]
Federated clustering enables multiple data sources to collaboratively cluster their data, maintaining decentralization and preserving privacy.
We introduce a novel federated clustering algorithm named Dynamically Weighted Federated k-means (DWF k-means) based on Lloyd's method for k-means clustering.
We conduct experiments on multiple datasets and data distribution settings to evaluate the performance of our algorithm in terms of clustering score, accuracy, and v-measure.
arXiv Detail & Related papers (2023-10-23T12:28:21Z) - Federated Deep Multi-View Clustering with Global Self-Supervision [51.639891178519136]
Federated multi-view clustering has the potential to learn a global clustering model from data distributed across multiple devices.
In this setting, label information is unknown and data privacy must be preserved.
We propose a novel federated deep multi-view clustering method that can mine complementary cluster structures from multiple clients.
arXiv Detail & Related papers (2023-09-24T17:07:01Z) - Timely Asynchronous Hierarchical Federated Learning: Age of Convergence [59.96266198512243]
We consider an asynchronous hierarchical federated learning setting with a client-edge-cloud framework.
The clients exchange the trained parameters with their corresponding edge servers, which update the locally aggregated model.
The goal of each client is to converge to the global model, while maintaining timeliness of the clients.
arXiv Detail & Related papers (2023-06-21T17:39:16Z) - Differentially Private Federated Clustering over Non-IID Data [59.611244450530315]
clustering clusters (FedC) problem aims to accurately partition unlabeled data samples distributed over massive clients into finite clients under the orchestration of a server.
We propose a novel FedC algorithm using differential privacy convergence technique, referred to as DP-Fed, in which partial participation and multiple clients are also considered.
Various attributes of the proposed DP-Fed are obtained through theoretical analyses of privacy protection, especially for the case of non-identically and independently distributed (non-i.i.d.) data.
arXiv Detail & Related papers (2023-01-03T05:38:43Z) - Privacy-Preserving Federated Deep Clustering based on GAN [12.256298398007848]
We present a novel approach to Federated Deep Clustering based on Generative Adversarial Networks (GANs)
Each client trains a local generative adversarial network (GAN) locally and uploads the synthetic data to the server.
The server applies a deep clustering network on the synthetic data to establish $k$ cluster centroids, which are then downloaded to the clients for cluster assignment.
arXiv Detail & Related papers (2022-11-30T13:20:11Z) - Optimizing Server-side Aggregation For Robust Federated Learning via
Subspace Training [80.03567604524268]
Non-IID data distribution across clients and poisoning attacks are two main challenges in real-world federated learning systems.
We propose SmartFL, a generic approach that optimize the server-side aggregation process.
We provide theoretical analyses of the convergence and generalization capacity for SmartFL.
arXiv Detail & Related papers (2022-11-10T13:20:56Z) - Federated clustering with GAN-based data synthesis [12.256298398007848]
Federated clustering (FC) is an extension of centralized clustering in federated settings.
We propose a new federated clustering framework, named synthetic data aided federated clustering (SDA-FC)
It trains generative adversarial network locally in each client and uploads the generated synthetic data to the server, where KM or FCM is performed on the synthetic data.
The synthetic data can make the model immune to the non-IID problem and enable us to capture the global similarity characteristics more effectively without sharing private data.
arXiv Detail & Related papers (2022-10-29T07:42:11Z) - Efficient Distribution Similarity Identification in Clustered Federated
Learning via Principal Angles Between Client Data Subspaces [59.33965805898736]
Clustered learning has been shown to produce promising results by grouping clients into clusters.
Existing FL algorithms are essentially trying to group clients together with similar distributions.
Prior FL algorithms attempt similarities indirectly during training.
arXiv Detail & Related papers (2022-09-21T17:37:54Z) - On the Convergence of Clustered Federated Learning [57.934295064030636]
In a federated learning system, the clients, e.g. mobile devices and organization participants, usually have different personal preferences or behavior patterns.
This paper proposes a novel weighted client-based clustered FL algorithm to leverage the client's group and each client in a unified optimization framework.
arXiv Detail & Related papers (2022-02-13T02:39:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.