When to Trust Aggregated Gradients: Addressing Negative Client Sampling
in Federated Learning
- URL: http://arxiv.org/abs/2301.10400v1
- Date: Wed, 25 Jan 2023 03:52:45 GMT
- Title: When to Trust Aggregated Gradients: Addressing Negative Client Sampling
in Federated Learning
- Authors: Wenkai Yang, Yankai Lin, Guangxiang Zhao, Peng Li, Jie Zhou, Xu Sun
- Abstract summary: We propose a novel learning rate adaptation mechanism to adjust the server learning rate for the aggregated gradient in each round.
We make theoretical deductions to find a meaningful and robust indicator that is positively related to the optimal server learning rate.
- Score: 41.51682329500003
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated Learning has become a widely-used framework which allows learning a
global model on decentralized local datasets under the condition of protecting
local data privacy. However, federated learning faces severe optimization
difficulty when training samples are not independently and identically
distributed (non-i.i.d.). In this paper, we point out that the client sampling
practice plays a decisive role in the aforementioned optimization difficulty.
We find that the negative client sampling will cause the merged data
distribution of currently sampled clients heavily inconsistent with that of all
available clients, and further make the aggregated gradient unreliable. To
address this issue, we propose a novel learning rate adaptation mechanism to
adaptively adjust the server learning rate for the aggregated gradient in each
round, according to the consistency between the merged data distribution of
currently sampled clients and that of all available clients. Specifically, we
make theoretical deductions to find a meaningful and robust indicator that is
positively related to the optimal server learning rate and can effectively
reflect the merged data distribution of sampled clients, and we utilize it for
the server learning rate adaptation. Extensive experiments on multiple image
and text classification tasks validate the great effectiveness of our method.
Related papers
- Federated Learning Clients Clustering with Adaptation to Data Drifts [27.974937897248132]
Federated Learning (FL) enables deep learning model training across edge devices.
In this paper, we introduce Fielding, a clustered FL framework that handles data drifts promptly with low overheads.
Our evaluations show that Fielding improves model final accuracy by 1.9%-5.9% and reaches target accuracies 1.16x-2.61x faster.
arXiv Detail & Related papers (2024-11-03T14:13:38Z) - LEFL: Low Entropy Client Sampling in Federated Learning [6.436397118145477]
Federated learning (FL) is a machine learning paradigm where multiple clients collaborate to optimize a single global model using their private data.
We propose LEFL, an alternative sampling strategy by performing a one-time clustering of clients based on their model's learned high-level features.
We show of sampled clients selected with this approach yield a low relative entropy with respect to the global data distribution.
arXiv Detail & Related papers (2023-12-29T01:44:20Z) - FedSampling: A Better Sampling Strategy for Federated Learning [81.85411484302952]
Federated learning (FL) is an important technique for learning models from decentralized data in a privacy-preserving way.
Existing FL methods usually uniformly sample clients for local model learning in each round.
We propose a novel data uniform sampling strategy for federated learning (FedSampling)
arXiv Detail & Related papers (2023-06-25T13:38:51Z) - Adaptive Self-Distillation for Minimizing Client Drift in Heterogeneous
Federated Learning [9.975023463908496]
Federated Learning (FL) is a machine learning paradigm that enables clients to jointly train a global model by aggregating the locally trained models without sharing any local training data.
We propose a novel regularization technique based on adaptive self-distillation (ASD) for training models on the client side.
Our regularization scheme adaptively adjusts to the client's training data based on the global model entropy and the client's label distribution.
arXiv Detail & Related papers (2023-05-31T07:00:42Z) - Federated Learning with Server Learning: Enhancing Performance for
Non-IID Data [5.070289965695956]
Federated Learning (FL) has emerged as a means of distributed learning using local data stored at clients with a coordinating server.
Recent studies showed that FL can suffer from poor performance and slower convergence when training data at clients are not independent and identically distributed.
Here we consider a new complementary approach to mitigating this performance degradation by allowing the server to perform auxiliary learning from a small dataset.
arXiv Detail & Related papers (2022-10-06T00:27:16Z) - Fed-CBS: A Heterogeneity-Aware Client Sampling Mechanism for Federated
Learning via Class-Imbalance Reduction [76.26710990597498]
We show that the class-imbalance of the grouped data from randomly selected clients can lead to significant performance degradation.
Based on our key observation, we design an efficient client sampling mechanism, i.e., Federated Class-balanced Sampling (Fed-CBS)
In particular, we propose a measure of class-imbalance and then employ homomorphic encryption to derive this measure in a privacy-preserving way.
arXiv Detail & Related papers (2022-09-30T05:42:56Z) - Straggler-Resilient Personalized Federated Learning [55.54344312542944]
Federated learning allows training models from samples distributed across a large network of clients while respecting privacy and communication restrictions.
We develop a novel algorithmic procedure with theoretical speedup guarantees that simultaneously handles two of these hurdles.
Our method relies on ideas from representation learning theory to find a global common representation using all clients' data and learn a user-specific set of parameters leading to a personalized solution for each client.
arXiv Detail & Related papers (2022-06-05T01:14:46Z) - Towards Fair Federated Learning with Zero-Shot Data Augmentation [123.37082242750866]
Federated learning has emerged as an important distributed learning paradigm, where a server aggregates a global model from many client-trained models while having no access to the client data.
We propose a novel federated learning system that employs zero-shot data augmentation on under-represented data to mitigate statistical heterogeneity and encourage more uniform accuracy performance across clients in federated networks.
We study two variants of this scheme, Fed-ZDAC (federated learning with zero-shot data augmentation at the clients) and Fed-ZDAS (federated learning with zero-shot data augmentation at the server).
arXiv Detail & Related papers (2021-04-27T18:23:54Z) - Exploiting Shared Representations for Personalized Federated Learning [54.65133770989836]
We propose a novel federated learning framework and algorithm for learning a shared data representation across clients and unique local heads for each client.
Our algorithm harnesses the distributed computational power across clients to perform many local-updates with respect to the low-dimensional local parameters for every update of the representation.
This result is of interest beyond federated learning to a broad class of problems in which we aim to learn a shared low-dimensional representation among data distributions.
arXiv Detail & Related papers (2021-02-14T05:36:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.