When Do Curricula Work in Federated Learning?
- URL: http://arxiv.org/abs/2212.12712v1
- Date: Sat, 24 Dec 2022 11:02:35 GMT
- Title: When Do Curricula Work in Federated Learning?
- Authors: Saeed Vahidian, Sreevatsank Kadaveru, Woonjoon Baek, Weijia Wang,
Vyacheslav Kungurtsev, Chen Chen, Mubarak Shah, and Bill Lin
- Abstract summary: We find that curriculum learning largely alleviates non-IIDness.
The more disparate the data distributions across clients the more they benefit from learning.
We propose a novel client selection technique that benefits from the real-world disparity in the clients.
- Score: 56.88941905240137
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An oft-cited open problem of federated learning is the existence of data
heterogeneity at the clients. One pathway to understanding the drastic accuracy
drop in federated learning is by scrutinizing the behavior of the clients' deep
models on data with different levels of "difficulty", which has been left
unaddressed. In this paper, we investigate a different and rarely studied
dimension of FL: ordered learning. Specifically, we aim to investigate how
ordered learning principles can contribute to alleviating the heterogeneity
effects in FL. We present theoretical analysis and conduct extensive empirical
studies on the efficacy of orderings spanning three kinds of learning:
curriculum, anti-curriculum, and random curriculum. We find that curriculum
learning largely alleviates non-IIDness. Interestingly, the more disparate the
data distributions across clients the more they benefit from ordered learning.
We provide analysis explaining this phenomenon, specifically indicating how
curriculum training appears to make the objective landscape progressively less
convex, suggesting fast converging iterations at the beginning of the training
procedure. We derive quantitative results of convergence for both convex and
nonconvex objectives by modeling the curriculum training on federated devices
as local SGD with locally biased stochastic gradients. Also, inspired by
ordered learning, we propose a novel client selection technique that benefits
from the real-world disparity in the clients. Our proposed approach to client
selection has a synergic effect when applied together with ordered learning in
FL.
Related papers
- SoK: Challenges and Opportunities in Federated Unlearning [32.0365189539138]
This SoK paper aims to take a deep look at the emphfederated unlearning literature, with the goal of identifying research trends and challenges in this emerging field.
arXiv Detail & Related papers (2024-03-04T19:35:08Z) - Federated Learning Can Find Friends That Are Advantageous [14.993730469216546]
In Federated Learning (FL), the distributed nature and heterogeneity of client data present both opportunities and challenges.
We introduce a novel algorithm that assigns adaptive aggregation weights to clients participating in FL training, identifying those with data distributions most conducive to a specific learning objective.
arXiv Detail & Related papers (2024-02-07T17:46:37Z) - FedLALR: Client-Specific Adaptive Learning Rates Achieve Linear Speedup
for Non-IID Data [54.81695390763957]
Federated learning is an emerging distributed machine learning method.
We propose a heterogeneous local variant of AMSGrad, named FedLALR, in which each client adjusts its learning rate.
We show that our client-specified auto-tuned learning rate scheduling can converge and achieve linear speedup with respect to the number of clients.
arXiv Detail & Related papers (2023-09-18T12:35:05Z) - UNIDEAL: Curriculum Knowledge Distillation Federated Learning [17.817181326740698]
Federated Learning (FL) has emerged as a promising approach to enable collaborative learning among multiple clients.
In this paper, we present UNI, a novel FL algorithm specifically designed to tackle the challenges of cross-domain scenarios.
Our results demonstrate that UNI achieves superior performance in terms of both model accuracy and communication efficiency.
arXiv Detail & Related papers (2023-09-16T11:30:29Z) - Collaborating Heterogeneous Natural Language Processing Tasks via
Federated Learning [55.99444047920231]
The proposed ATC framework achieves significant improvements compared with various baseline methods.
We conduct extensive experiments on six widely-used datasets covering both Natural Language Understanding (NLU) and Natural Language Generation (NLG) tasks.
arXiv Detail & Related papers (2022-12-12T09:27:50Z) - Addressing Client Drift in Federated Continual Learning with Adaptive
Optimization [10.303676184878896]
We outline a framework for performing Federated Continual Learning (FCL) by using NetTailor as a candidate continual learning approach.
We show that adaptive federated optimization can reduce the adverse impact of client drift and showcase its effectiveness on CIFAR100, MiniImagenet, and Decathlon benchmarks.
arXiv Detail & Related papers (2022-03-24T20:00:03Z) - Exploiting Shared Representations for Personalized Federated Learning [54.65133770989836]
We propose a novel federated learning framework and algorithm for learning a shared data representation across clients and unique local heads for each client.
Our algorithm harnesses the distributed computational power across clients to perform many local-updates with respect to the low-dimensional local parameters for every update of the representation.
This result is of interest beyond federated learning to a broad class of problems in which we aim to learn a shared low-dimensional representation among data distributions.
arXiv Detail & Related papers (2021-02-14T05:36:25Z) - Toward Understanding the Influence of Individual Clients in Federated
Learning [52.07734799278535]
Federated learning allows clients to jointly train a global model without sending their private data to a central server.
We defined a new notion called em-Influence, quantify this influence over parameters, and proposed an effective efficient model to estimate this metric.
arXiv Detail & Related papers (2020-12-20T14:34:36Z) - Federated Continual Learning with Weighted Inter-client Transfer [79.93004004545736]
We propose a novel federated continual learning framework, Federated Weighted Inter-client Transfer (FedWeIT)
FedWeIT decomposes the network weights into global federated parameters and sparse task-specific parameters, and each client receives selective knowledge from other clients.
We validate our FedWeIT against existing federated learning and continual learning methods, and our model significantly outperforms them with a large reduction in the communication cost.
arXiv Detail & Related papers (2020-03-06T13:33:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.