An Efficient Virtual Data Generation Method for Reducing Communication
in Federated Learning
- URL: http://arxiv.org/abs/2306.12088v3
- Date: Thu, 29 Jun 2023 08:44:48 GMT
- Title: An Efficient Virtual Data Generation Method for Reducing Communication
in Federated Learning
- Authors: Cheng Yang, Xue Yang, Dongxian Wu, Xiaohu Tang
- Abstract summary: A few classical schemes assume the server can extract the auxiliary information about training data of the participants from the local models to construct a central dummy dataset.
The server uses the dummy dataset to finetune aggregated global model to achieve the target test accuracy in fewer communication rounds.
In this paper, we summarize the above solutions into a data-based communication-efficient FL framework.
- Score: 34.39250699866746
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Communication overhead is one of the major challenges in Federated
Learning(FL). A few classical schemes assume the server can extract the
auxiliary information about training data of the participants from the local
models to construct a central dummy dataset. The server uses the dummy dataset
to finetune aggregated global model to achieve the target test accuracy in
fewer communication rounds. In this paper, we summarize the above solutions
into a data-based communication-efficient FL framework. The key of the proposed
framework is to design an efficient extraction module(EM) which ensures the
dummy dataset has a positive effect on finetuning aggregated global model.
Different from the existing methods that use generator to design EM, our
proposed method, FedINIBoost borrows the idea of gradient match to construct
EM. Specifically, FedINIBoost builds a proxy dataset of the real dataset in two
steps for each participant at each communication round. Then the server
aggregates all the proxy datasets to form a central dummy dataset, which is
used to finetune aggregated global model. Extensive experiments verify the
superiority of our method compared with the existing classical method, FedAVG,
FedProx, Moon and FedFTG. Moreover, FedINIBoost plays a significant role in
finetuning the performance of aggregated global model at the initial stage of
FL.
Related papers
- FedMAP: Unlocking Potential in Personalized Federated Learning through Bi-Level MAP Optimization [11.040916982022978]
Federated Learning (FL) enables collaborative training of machine learning models on decentralized data.
Data across clients often differs significantly due to class imbalance, feature distribution skew, sample size imbalance, and other phenomena.
We propose a novel Bayesian PFL framework using bi-level optimization to tackle the data heterogeneity challenges.
arXiv Detail & Related papers (2024-05-29T11:28:06Z) - An Aggregation-Free Federated Learning for Tackling Data Heterogeneity [50.44021981013037]
Federated Learning (FL) relies on the effectiveness of utilizing knowledge from distributed datasets.
Traditional FL methods adopt an aggregate-then-adapt framework, where clients update local models based on a global model aggregated by the server from the previous training round.
We introduce FedAF, a novel aggregation-free FL algorithm.
arXiv Detail & Related papers (2024-04-29T05:55:23Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Fake It Till Make It: Federated Learning with Consensus-Oriented
Generation [52.82176415223988]
We propose federated learning with consensus-oriented generation (FedCOG)
FedCOG consists of two key components at the client side: complementary data generation and knowledge-distillation-based model training.
Experiments on classical and real-world FL datasets show that FedCOG consistently outperforms state-of-the-art methods.
arXiv Detail & Related papers (2023-12-10T18:49:59Z) - FedCME: Client Matching and Classifier Exchanging to Handle Data
Heterogeneity in Federated Learning [5.21877373352943]
Data heterogeneity across clients is one of the key challenges in Federated Learning (FL)
We propose a novel FL framework named FedCME by client matching and classifier exchanging.
Experimental results demonstrate that FedCME performs better than FedAvg, FedProx, MOON and FedRS on popular federated learning benchmarks.
arXiv Detail & Related papers (2023-07-17T15:40:45Z) - Optimizing Server-side Aggregation For Robust Federated Learning via
Subspace Training [80.03567604524268]
Non-IID data distribution across clients and poisoning attacks are two main challenges in real-world federated learning systems.
We propose SmartFL, a generic approach that optimize the server-side aggregation process.
We provide theoretical analyses of the convergence and generalization capacity for SmartFL.
arXiv Detail & Related papers (2022-11-10T13:20:56Z) - FedDM: Iterative Distribution Matching for Communication-Efficient
Federated Learning [87.08902493524556]
Federated learning(FL) has recently attracted increasing attention from academia and industry.
We propose FedDM to build the global training objective from multiple local surrogate functions.
In detail, we construct synthetic sets of data on each client to locally match the loss landscape from original data.
arXiv Detail & Related papers (2022-07-20T04:55:18Z) - Toward Understanding the Influence of Individual Clients in Federated
Learning [52.07734799278535]
Federated learning allows clients to jointly train a global model without sending their private data to a central server.
We defined a new notion called em-Influence, quantify this influence over parameters, and proposed an effective efficient model to estimate this metric.
arXiv Detail & Related papers (2020-12-20T14:34:36Z) - Analysis and Optimal Edge Assignment For Hierarchical Federated Learning
on Non-IID Data [43.32085029569374]
Federated learning algorithms aim to leverage distributed and diverse data stored at users' devices to learn a global phenomena.
In the cases where the participants' data are strongly skewed (i.e., non-IID), the local models can overfit local data, leading to low performing global model.
We propose a hierarchical learning system that performs Federated Gradient Descent on the user-edge layer and Federated Averaging on the edge-cloud layer.
arXiv Detail & Related papers (2020-12-10T12:18:13Z) - FedFMC: Sequential Efficient Federated Learning on Non-iid Data [0.0]
FedFMC (Fork-Consolidate-Merge) is a method that forks devices into updating different global models then merges and consolidates separate models into one.
We show that FedFMC substantially improves upon earlier approaches to non-iid data in the federated learning context without using a globally shared subset of data nor increase communication costs.
arXiv Detail & Related papers (2020-06-19T02:36:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.