FedBN: Federated Learning on Non-IID Features via Local Batch
Normalization
- URL: http://arxiv.org/abs/2102.07623v1
- Date: Mon, 15 Feb 2021 16:04:10 GMT
- Title: FedBN: Federated Learning on Non-IID Features via Local Batch
Normalization
- Authors: Xiaoxiao Li, Meirui Jiang, Xiaofei Zhang, Michael Kamp, Qi Dou
- Abstract summary: The emerging paradigm of federated learning (FL) strives to enable collaborative training of deep models on the network edge without centrally aggregating raw data.
We propose an effective method that uses local batch normalization to alleviate the feature shift before averaging models.
The resulting scheme, called FedBN, outperforms both classical FedAvg and the state-of-the-art for non-iid data.
- Score: 23.519212374186232
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The emerging paradigm of federated learning (FL) strives to enable
collaborative training of deep models on the network edge without centrally
aggregating raw data and hence improving data privacy. In most cases, the
assumption of independent and identically distributed samples across local
clients does not hold for federated learning setups. Under this setting, neural
network training performance may vary significantly according to the data
distribution and even hurt training convergence. Most of the previous work has
focused on a difference in the distribution of labels or client shifts. Unlike
those settings, we address an important problem of FL, e.g., different
scanners/sensors in medical imaging, different scenery distribution in
autonomous driving (highway vs. city), where local clients store examples with
different distributions compared to other clients, which we denote as feature
shift non-iid. In this work, we propose an effective method that uses local
batch normalization to alleviate the feature shift before averaging models. The
resulting scheme, called FedBN, outperforms both classical FedAvg, as well as
the state-of-the-art for non-iid data (FedProx) on our extensive experiments.
These empirical results are supported by a convergence analysis that shows in a
simplified setting that FedBN has a faster convergence rate than FedAvg. Code
is available at https://github.com/med-air/FedBN.
Related papers
- FedImpro: Measuring and Improving Client Update in Federated Learning [77.68805026788836]
Federated Learning (FL) models often experience client drift caused by heterogeneous data.
We present an alternative perspective on client drift and aim to mitigate it by generating improved local models.
arXiv Detail & Related papers (2024-02-10T18:14:57Z) - Leveraging Function Space Aggregation for Federated Learning at Scale [20.866482460590973]
We propose a new algorithm, FedFish, that aggregates local approximations to the functions learned by clients.
We evaluate FedFish on realistic, large-scale cross-device benchmarks.
arXiv Detail & Related papers (2023-11-17T02:37:10Z) - A Simple Data Augmentation for Feature Distribution Skewed Federated Learning [47.27053883247425]
Federated Learning (FL) facilitates collaborative learning among multiple clients in a distributed manner.
FL's performance degrades with non-Independent and Identically Distributed (non-IID) data.
We propose FedRDN, which randomly injects the statistical information of the local distribution from the entire federation into the client's data.
Our FedRDN is a plug-and-play component, which can be seamlessly integrated into the data augmentation flow with only a few lines of code.
arXiv Detail & Related papers (2023-06-14T05:46:52Z) - Benchmarking FedAvg and FedCurv for Image Classification Tasks [1.376408511310322]
This paper focuses on the problem of statistical heterogeneity of the data in the same federated network.
Several Federated Learning algorithms, such as FedAvg, FedProx and Federated Curvature (FedCurv) have already been proposed.
As a side product of this work, we release the non-IID version of the datasets we used so to facilitate further comparisons from the FL community.
arXiv Detail & Related papers (2023-03-31T10:13:01Z) - FedFA: Federated Feature Augmentation [25.130087374092383]
Federated learning allows multiple parties to collaboratively train deep models without exchanging raw data.
The primary goal of this paper is to develop a robust federated learning algorithm to address feature shift in clients' samples.
We propose FedFA to tackle federated learning from a distinct perspective of federated feature augmentation.
arXiv Detail & Related papers (2023-01-30T15:39:55Z) - FedSkip: Combatting Statistical Heterogeneity with Federated Skip
Aggregation [95.85026305874824]
We introduce a data-driven approach called FedSkip to improve the client optima by periodically skipping federated averaging and scattering local models to the cross devices.
We conduct extensive experiments on a range of datasets to demonstrate that FedSkip achieves much higher accuracy, better aggregation efficiency and competing communication efficiency.
arXiv Detail & Related papers (2022-12-14T13:57:01Z) - FedAvg with Fine Tuning: Local Updates Lead to Representation Learning [54.65133770989836]
Federated Averaging (FedAvg) algorithm consists of alternating between a few local gradient updates at client nodes, followed by a model averaging update at the server.
We show that the reason behind generalizability of the FedAvg's output is its power in learning the common data representation among the clients' tasks.
We also provide empirical evidence demonstrating FedAvg's representation learning ability in federated image classification with heterogeneous data.
arXiv Detail & Related papers (2022-05-27T00:55:24Z) - FedDC: Federated Learning with Non-IID Data via Local Drift Decoupling
and Correction [48.85303253333453]
Federated learning (FL) allows multiple clients to collectively train a high-performance global model without sharing their private data.
We propose a novel federated learning algorithm with local drift decoupling and correction (FedDC)
Our FedDC only introduces lightweight modifications in the local training phase, in which each client utilizes an auxiliary local drift variable to track the gap between the local model parameter and the global model parameters.
Experiment results and analysis demonstrate that FedDC yields expediting convergence and better performance on various image classification tasks.
arXiv Detail & Related papers (2022-03-22T14:06:26Z) - Acceleration of Federated Learning with Alleviated Forgetting in Local
Training [61.231021417674235]
Federated learning (FL) enables distributed optimization of machine learning models while protecting privacy.
We propose FedReg, an algorithm to accelerate FL with alleviated knowledge forgetting in the local training stage.
Our experiments demonstrate that FedReg not only significantly improves the convergence rate of FL, especially when the neural network architecture is deep.
arXiv Detail & Related papers (2022-03-05T02:31:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.