Critical Learning Periods in Federated Learning
- URL: http://arxiv.org/abs/2109.05613v1
- Date: Sun, 12 Sep 2021 21:06:07 GMT
- Title: Critical Learning Periods in Federated Learning
- Authors: Gang Yan, Hao Wang, Jian Li
- Abstract summary: Federated learning (FL) is a popular technique to train machine learning (ML) models with decentralized data.
We show that the final test accuracy of FL is dramatically affected by the early phase of the training process.
- Score: 11.138980572551066
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning (FL) is a popular technique to train machine learning (ML)
models with decentralized data. Extensive works have studied the performance of
the global model; however, it is still unclear how the training process affects
the final test accuracy. Exacerbating this problem is the fact that FL
executions differ significantly from traditional ML with heterogeneous data
characteristics across clients, involving more hyperparameters. In this work,
we show that the final test accuracy of FL is dramatically affected by the
early phase of the training process, i.e., FL exhibits critical learning
periods, in which small gradient errors can have irrecoverable impact on the
final test accuracy. To further explain this phenomenon, we generalize the
trace of the Fisher Information Matrix (FIM) to FL and define a new notion
called FedFIM, a quantity reflecting the local curvature of each clients from
the beginning of the training in FL. Our findings suggest that the {\em initial
learning phase} plays a critical role in understanding the FL performance. This
is in contrast to many existing works which generally do not connect the final
accuracy of FL to the early phase training. Finally, seizing critical learning
periods in FL is of independent interest and could be useful for other problems
such as the choices of hyperparameters such as the number of client selected
per round, batch size, and more, so as to improve the performance of FL
training and testing.
Related papers
- Can We Theoretically Quantify the Impacts of Local Updates on the Generalization Performance of Federated Learning? [50.03434441234569]
Federated Learning (FL) has gained significant popularity due to its effectiveness in training machine learning models across diverse sites without requiring direct data sharing.
While various algorithms have shown that FL with local updates is a communication-efficient distributed learning framework, the generalization performance of FL with local updates has received comparatively less attention.
arXiv Detail & Related papers (2024-09-05T19:00:18Z) - Rethinking the Starting Point: Collaborative Pre-Training for Federated Downstream Tasks [21.842345900168525]
CoPreFL is a model-agnostic meta-learning (MAML) procedure that tailors the global model to closely mimic heterogeneous and unseen FL scenarios.
Our MAML procedure incorporates performance variance into the meta-objective function, balancing performance across clients.
We demonstrate that CoPreFL obtains significant improvements in both average accuracy and variance across arbitrary downstream FL tasks.
arXiv Detail & Related papers (2024-02-03T17:58:43Z) - A Survey on Efficient Federated Learning Methods for Foundation Model Training [62.473245910234304]
Federated Learning (FL) has become an established technique to facilitate privacy-preserving collaborative training across a multitude of clients.
In the wake of Foundation Models (FM), the reality is different for many deep learning applications.
We discuss the benefits and drawbacks of parameter-efficient fine-tuning (PEFT) for FL applications.
arXiv Detail & Related papers (2024-01-09T10:22:23Z) - A Comprehensive View of Personalized Federated Learning on Heterogeneous Clinical Datasets [0.4926316920996346]
Federated learning (FL) is a key approach to overcoming the data silos that so frequently obstruct the training and deployment of machine-learning models in clinical settings.
This work contributes to a growing body of FL research specifically focused on clinical applications along three important directions.
arXiv Detail & Related papers (2023-09-28T20:12:17Z) - Understanding How Consistency Works in Federated Learning via Stage-wise
Relaxed Initialization [84.42306265220274]
Federated learning (FL) is a distributed paradigm that coordinates massive local clients to collaboratively train a global model.
Previous works have implicitly studied that FL suffers from the client-drift'' problem, which is caused by the inconsistent optimum across local clients.
To alleviate the negative impact of the client drift'' and explore its substance in FL, we first design an efficient FL algorithm textitFedInit.
arXiv Detail & Related papers (2023-06-09T06:55:15Z) - Automated Federated Learning in Mobile Edge Networks -- Fast Adaptation
and Convergence [83.58839320635956]
Federated Learning (FL) can be used in mobile edge networks to train machine learning models in a distributed manner.
Recent FL has been interpreted within a Model-Agnostic Meta-Learning (MAML) framework, which brings FL significant advantages in fast adaptation and convergence over heterogeneous datasets.
This paper addresses how much benefit MAML brings to FL and how to maximize such benefit over mobile edge networks.
arXiv Detail & Related papers (2023-03-23T02:42:10Z) - FL Games: A Federated Learning Framework for Distribution Shifts [71.98708418753786]
Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server.
We propose FL GAMES, a game-theoretic framework for federated learning that learns causal features that are invariant across clients.
arXiv Detail & Related papers (2022-10-31T22:59:03Z) - On the Importance and Applicability of Pre-Training for Federated
Learning [28.238484580662785]
We conduct a systematic study to explore pre-training for federated learning.
We find that pre-training can improve FL, but also close its accuracy gap to the counterpart centralized learning.
We conclude our paper with an attempt to understand the effect of pre-training on FL.
arXiv Detail & Related papers (2022-06-23T06:02:33Z) - FL Games: A federated learning framework for distribution shifts [71.98708418753786]
Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server.
We propose FL Games, a game-theoretic framework for federated learning for learning causal features that are invariant across clients.
arXiv Detail & Related papers (2022-05-23T07:51:45Z) - Federated Unlearning [24.60965999954735]
Federated learning (FL) has emerged as a promising distributed machine learning (ML) paradigm.
Practical needs of the "right to be forgotten" and countering data poisoning attacks call for efficient techniques that can remove, or unlearn, specific training data from the trained FL model.
We present FedEraser, the first federated unlearning methodology that can eliminate the influence of a federated client's data on the global FL model.
arXiv Detail & Related papers (2020-12-27T08:54:37Z) - Addressing Class Imbalance in Federated Learning [10.970632986559547]
Federated learning (FL) is a promising approach for training decentralized data located on local client devices.
We propose a monitoring scheme that can infer the composition of training data for each FL round, and design a new loss function -- textbfRatio Loss to mitigate the impact.
arXiv Detail & Related papers (2020-08-14T07:28:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.