FedCBO: Reaching Group Consensus in Clustered Federated Learning through
Consensus-based Optimization
- URL: http://arxiv.org/abs/2305.02894v1
- Date: Thu, 4 May 2023 15:02:09 GMT
- Title: FedCBO: Reaching Group Consensus in Clustered Federated Learning through
Consensus-based Optimization
- Authors: Jose A. Carrillo, Nicolas Garcia Trillos, Sixu Li, Yuhua Zhu
- Abstract summary: Federated learning seeks to integrate the training learning models from multiple users, each user having their own data set, in a way that is sensitive to data privacy and to communication loss constraints.
In this paper, we propose a novel solution to a global, clustered problem of federated learning that is inspired by ideas in consensus-based optimization (CBO)
Our new CBO-type method is based on a system of interacting particles that is oblivious to group.
- Score: 1.911678487931003
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning is an important framework in modern machine learning that
seeks to integrate the training of learning models from multiple users, each
user having their own local data set, in a way that is sensitive to data
privacy and to communication loss constraints. In clustered federated learning,
one assumes an additional unknown group structure among users, and the goal is
to train models that are useful for each group, rather than simply training a
single global model for all users. In this paper, we propose a novel solution
to the problem of clustered federated learning that is inspired by ideas in
consensus-based optimization (CBO). Our new CBO-type method is based on a
system of interacting particles that is oblivious to group memberships. Our
model is motivated by rigorous mathematical reasoning, including a mean field
analysis describing the large number of particles limit of our particle system,
as well as convergence guarantees for the simultaneous global optimization of
general non-convex objective functions (corresponding to the loss functions of
each cluster of users) in the mean-field regime. Experimental results
demonstrate the efficacy of our FedCBO algorithm compared to other
state-of-the-art methods and help validate our methodological and theoretical
work.
Related papers
- Guaranteeing Data Privacy in Federated Unlearning with Dynamic User Participation [21.07328631033828]
Federated Unlearning (FU) can eliminate influences of Federated Learning (FL) users' data from trained global FL models.
A straightforward FU method involves removing the unlearned users and subsequently retraining a new global FL model from scratch with all remaining users.
We propose a privacy-preserving FU framework, aimed at ensuring privacy while effectively managing dynamic user participation.
arXiv Detail & Related papers (2024-06-03T03:39:07Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Efficient Cluster Selection for Personalized Federated Learning: A
Multi-Armed Bandit Approach [2.5477011559292175]
Federated learning (FL) offers a decentralized training approach for machine learning models, prioritizing data privacy.
In this paper, we introduce a dynamic Upper Confidence Bound (dUCB) algorithm inspired by the multi-armed bandit (MAB) approach.
arXiv Detail & Related papers (2023-10-29T16:46:50Z) - Federated cINN Clustering for Accurate Clustered Federated Learning [33.72494731516968]
Federated Learning (FL) presents an innovative approach to privacy-preserving distributed machine learning.
We propose the Federated cINN Clustering Algorithm (FCCA) to robustly cluster clients into different groups.
arXiv Detail & Related papers (2023-09-04T10:47:52Z) - Tackling Computational Heterogeneity in FL: A Few Theoretical Insights [68.8204255655161]
We introduce and analyse a novel aggregation framework that allows for formalizing and tackling computational heterogeneous data.
Proposed aggregation algorithms are extensively analyzed from a theoretical, and an experimental prospective.
arXiv Detail & Related papers (2023-07-12T16:28:21Z) - Personalizing Federated Learning with Over-the-Air Computations [84.8089761800994]
Federated edge learning is a promising technology to deploy intelligence at the edge of wireless networks in a privacy-preserving manner.
Under such a setting, multiple clients collaboratively train a global generic model under the coordination of an edge server.
This paper presents a distributed training paradigm that employs analog over-the-air computation to address the communication bottleneck.
arXiv Detail & Related papers (2023-02-24T08:41:19Z) - Proof of Swarm Based Ensemble Learning for Federated Learning
Applications [3.2536767864585663]
In federated learning it is not feasible to apply centralised ensemble learning directly due to privacy concerns.
Most distributed consensus algorithms, such as Byzantine fault tolerance (BFT), do not normally perform well in such applications.
We propose PoSw, a novel distributed consensus algorithm for ensemble learning in a federated setting.
arXiv Detail & Related papers (2022-12-28T13:53:34Z) - DRFLM: Distributionally Robust Federated Learning with Inter-client
Noise via Local Mixup [58.894901088797376]
federated learning has emerged as a promising approach for training a global model using data from multiple organizations without leaking their raw data.
We propose a general framework to solve the above two challenges simultaneously.
We provide comprehensive theoretical analysis including robustness analysis, convergence analysis, and generalization ability.
arXiv Detail & Related papers (2022-04-16T08:08:29Z) - On the Convergence of Clustered Federated Learning [57.934295064030636]
In a federated learning system, the clients, e.g. mobile devices and organization participants, usually have different personal preferences or behavior patterns.
This paper proposes a novel weighted client-based clustered FL algorithm to leverage the client's group and each client in a unified optimization framework.
arXiv Detail & Related papers (2022-02-13T02:39:19Z) - An Efficient Framework for Clustered Federated Learning [26.24231986590374]
We address the problem of federated learning (FL) where users are distributed into clusters.
We propose the Iterative Federated Clustering Algorithm (IFCA)
We show that our algorithm is efficient in non- partitioned problems such as neural networks.
arXiv Detail & Related papers (2020-06-07T08:48:59Z) - Multi-Center Federated Learning [62.57229809407692]
This paper proposes a novel multi-center aggregation mechanism for federated learning.
It learns multiple global models from the non-IID user data and simultaneously derives the optimal matching between users and centers.
Our experimental results on benchmark datasets show that our method outperforms several popular federated learning methods.
arXiv Detail & Related papers (2020-05-03T09:14:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.