Fed-ensemble: Improving Generalization through Model Ensembling in
Federated Learning
- URL: http://arxiv.org/abs/2107.10663v1
- Date: Wed, 21 Jul 2021 14:40:14 GMT
- Title: Fed-ensemble: Improving Generalization through Model Ensembling in
Federated Learning
- Authors: Naichen Shi, Fan Lai, Raed Al Kontar, Mosharaf Chowdhury
- Abstract summary: Fed-ensemble brings model ensembling to federated learning (FL)
Fed-ensemble can be readily utilized within established FL methods.
- Score: 5.882234707363695
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper we propose Fed-ensemble: a simple approach that bringsmodel
ensembling to federated learning (FL). Instead of aggregating localmodels to
update a single global model, Fed-ensemble uses random permutations to update a
group of K models and then obtains predictions through model averaging.
Fed-ensemble can be readily utilized within established FL methods and does not
impose a computational overhead as it only requires one of the K models to be
sent to a client in each communication round. Theoretically, we show that
predictions on newdata from all K models belong to the same predictive
posterior distribution under a neural tangent kernel regime. This result in
turn sheds light onthe generalization advantages of model averaging. We also
illustrate thatFed-ensemble has an elegant Bayesian interpretation. Empirical
results show that our model has superior performance over several FL
algorithms,on a wide range of data sets, and excels in heterogeneous settings
often encountered in FL applications.
Related papers
- Aggregation Weighting of Federated Learning via Generalization Bound
Estimation [65.8630966842025]
Federated Learning (FL) typically aggregates client model parameters using a weighting approach determined by sample proportions.
We replace the aforementioned weighting method with a new strategy that considers the generalization bounds of each local model.
arXiv Detail & Related papers (2023-11-10T08:50:28Z) - Federated Variational Inference: Towards Improved Personalization and
Generalization [2.37589914835055]
We study personalization and generalization in stateless cross-device federated learning setups.
We first propose a hierarchical generative model and formalize it using Bayesian Inference.
We then approximate this process using Variational Inference to train our model efficiently.
We evaluate our model on FEMNIST and CIFAR-100 image classification and show that FedVI beats the state-of-the-art on both tasks.
arXiv Detail & Related papers (2023-05-23T04:28:07Z) - FedHB: Hierarchical Bayesian Federated Learning [11.936836827864095]
We propose a novel hierarchical Bayesian approach to Federated Learning (FL)
Our model reasonably describes the generative process of clients' local data via hierarchical Bayesian modeling.
We show that our block-coordinate FL algorithm converges to an optimum of the objective at the rate of $O(sqrtt)$.
arXiv Detail & Related papers (2023-05-08T18:21:41Z) - Learning from aggregated data with a maximum entropy model [73.63512438583375]
We show how a new model, similar to a logistic regression, may be learned from aggregated data only by approximating the unobserved feature distribution with a maximum entropy hypothesis.
We present empirical evidence on several public datasets that the model learned this way can achieve performances comparable to those of a logistic model trained with the full unaggregated data.
arXiv Detail & Related papers (2022-10-05T09:17:27Z) - FedAvg with Fine Tuning: Local Updates Lead to Representation Learning [54.65133770989836]
Federated Averaging (FedAvg) algorithm consists of alternating between a few local gradient updates at client nodes, followed by a model averaging update at the server.
We show that the reason behind generalizability of the FedAvg's output is its power in learning the common data representation among the clients' tasks.
We also provide empirical evidence demonstrating FedAvg's representation learning ability in federated image classification with heterogeneous data.
arXiv Detail & Related papers (2022-05-27T00:55:24Z) - Federated Submodel Optimization for Hot and Cold Data Features [41.99190452773989]
We study practical data characteristics underlying federated learning, where non-i.i.d. data from clients have sparse features, and a certain client's local data normally involves only a small part of the full model.
Due to data sparsity, the classical federated averaging (FedAvg) algorithm or its variants will be severely slowed down.
We propose federated submodel averaging (FedSubAvg), ensuring that the expectation of the global update of each model parameter is equal to the average of the local updates of the clients.
arXiv Detail & Related papers (2021-09-16T03:54:54Z) - FedBE: Making Bayesian Model Ensemble Applicable to Federated Learning [23.726336635748783]
Federated learning aims to collaboratively train a strong global model by accessing users' locally trained models but not their own data.
A crucial step is therefore to aggregate local models into a global model, which has been shown challenging when users have non-i.i.d. data.
We propose a novel aggregation algorithm named FedBE, which takes a Bayesian inference perspective by sampling higher-quality global models.
arXiv Detail & Related papers (2020-09-04T01:18:25Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z) - UVeQFed: Universal Vector Quantization for Federated Learning [179.06583469293386]
Federated learning (FL) is an emerging approach to train such learning models without requiring the users to share their possibly private labeled data.
In FL, each user trains its copy of the learning model locally. The server then collects the individual updates and aggregates them into a global model.
We show that combining universal vector quantization methods with FL yields a decentralized training system in which the compression of the trained models induces only a minimum distortion.
arXiv Detail & Related papers (2020-06-05T07:10:22Z) - Federated learning with hierarchical clustering of local updates to
improve training on non-IID data [3.3517146652431378]
We show that learning a single joint model is often not optimal in the presence of certain types of non-iid data.
We present a modification to FL by introducing a hierarchical clustering step (FL+HC)
We show how FL+HC allows model training to converge in fewer communication rounds compared to FL without clustering.
arXiv Detail & Related papers (2020-04-24T15:16:01Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.