DisPFL: Towards Communication-Efficient Personalized Federated Learning
via Decentralized Sparse Training
- URL: http://arxiv.org/abs/2206.00187v1
- Date: Wed, 1 Jun 2022 02:20:57 GMT
- Title: DisPFL: Towards Communication-Efficient Personalized Federated Learning
via Decentralized Sparse Training
- Authors: Rong Dai, Li Shen, Fengxiang He, Xinmei Tian, Dacheng Tao
- Abstract summary: We propose a novel personalized federated learning framework in a decentralized (peer-to-peer) communication protocol named Dis-PFL.
Dis-PFL employs personalized sparse masks to customize sparse local models on the edge.
We demonstrate that our method can easily adapt to heterogeneous local clients with varying computation complexities.
- Score: 84.81043932706375
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Personalized federated learning is proposed to handle the data heterogeneity
problem amongst clients by learning dedicated tailored local models for each
user. However, existing works are often built in a centralized way, leading to
high communication pressure and high vulnerability when a failure or an attack
on the central server occurs. In this work, we propose a novel personalized
federated learning framework in a decentralized (peer-to-peer) communication
protocol named Dis-PFL, which employs personalized sparse masks to customize
sparse local models on the edge. To further save the communication and
computation cost, we propose a decentralized sparse training technique, which
means that each local model in Dis-PFL only maintains a fixed number of active
parameters throughout the whole local training and peer-to-peer communication
process. Comprehensive experiments demonstrate that Dis-PFL significantly saves
the communication bottleneck for the busiest node among all clients and, at the
same time, achieves higher model accuracy with less computation cost and
communication rounds. Furthermore, we demonstrate that our method can easily
adapt to heterogeneous local clients with varying computation complexities and
achieves better personalized performances.
Related papers
- Prune at the Clients, Not the Server: Accelerated Sparse Training in Federated Learning [56.21666819468249]
Resource constraints of clients and communication costs pose major problems for training large models in Federated Learning.
We introduce Sparse-ProxSkip, which combines training and acceleration in a sparse setting.
We demonstrate the good performance of Sparse-ProxSkip in extensive experiments.
arXiv Detail & Related papers (2024-05-31T05:21:12Z) - Decentralized Directed Collaboration for Personalized Federated Learning [39.29794569421094]
We concentrate on the Decentralized Personalized Learning (DPFL) that performs distributed training model computation.
We propose a directed collaboration framework by incorporating textbfDecentralized textbfFederated textbfPartial textbfGradient textbfPedGP.
arXiv Detail & Related papers (2024-05-28T06:52:19Z) - FedComLoc: Communication-Efficient Distributed Training of Sparse and Quantized Models [56.21666819468249]
Federated Learning (FL) has garnered increasing attention due to its unique characteristic of allowing heterogeneous clients to process their private data locally and interact with a central server.
We introduce FedComLoc, integrating practical and effective compression into emphScaffnew to further enhance communication efficiency.
arXiv Detail & Related papers (2024-03-14T22:29:59Z) - Scheduling and Communication Schemes for Decentralized Federated
Learning [0.31410859223862103]
A decentralized federated learning (DFL) model with the gradient descent (SGD) algorithm has been introduced.
Three scheduling policies for DFL have been proposed for communications between the clients and the parallel servers.
Results show that the proposed scheduling polices have an impact both on the speed of convergence and in the final global model.
arXiv Detail & Related papers (2023-11-27T17:35:28Z) - Tunable Soft Prompts are Messengers in Federated Learning [55.924749085481544]
Federated learning (FL) enables multiple participants to collaboratively train machine learning models using decentralized data sources.
The lack of model privacy protection in FL becomes an unneglectable challenge.
We propose a novel FL training approach that accomplishes information exchange among participants via tunable soft prompts.
arXiv Detail & Related papers (2023-11-12T11:01:10Z) - Towards More Suitable Personalization in Federated Learning via
Decentralized Partial Model Training [67.67045085186797]
Almost all existing systems have to face large communication burdens if the central FL server fails.
It personalizes the "right" in the deep models by alternately updating the shared and personal parameters.
To further promote the shared parameters aggregation process, we propose DFed integrating the local Sharpness Miniization.
arXiv Detail & Related papers (2023-05-24T13:52:18Z) - Explicit Personalization and Local Training: Double Communication
Acceleration in Federated Learning [7.691755449724637]
A common strategy to curtail communication costs is Local Training, which consists in performing multiple local gradient descent steps between successive communication rounds.
We introduce Scafflix, a novel algorithm that efficiently integrates explicit personalization with local training.
arXiv Detail & Related papers (2023-05-22T15:58:01Z) - SalientGrads: Sparse Models for Communication Efficient and Data Aware
Distributed Federated Training [1.0413504599164103]
Federated learning (FL) enables the training of a model leveraging decentralized data in client sites while preserving privacy by not collecting data.
One of the significant challenges of FL is limited computation and low communication bandwidth in resource limited edge client nodes.
We propose Salient Grads, which simplifies the process of sparse training by choosing a data aware subnetwork before training.
arXiv Detail & Related papers (2023-04-15T06:46:37Z) - Personalizing Federated Learning with Over-the-Air Computations [84.8089761800994]
Federated edge learning is a promising technology to deploy intelligence at the edge of wireless networks in a privacy-preserving manner.
Under such a setting, multiple clients collaboratively train a global generic model under the coordination of an edge server.
This paper presents a distributed training paradigm that employs analog over-the-air computation to address the communication bottleneck.
arXiv Detail & Related papers (2023-02-24T08:41:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.