FedSDD: Scalable and Diversity-enhanced Distillation for Model
Aggregation in Federated Learning
- URL: http://arxiv.org/abs/2312.17029v1
- Date: Thu, 28 Dec 2023 14:10:00 GMT
- Title: FedSDD: Scalable and Diversity-enhanced Distillation for Model
Aggregation in Federated Learning
- Authors: Ho Man Kwan, Shenghui Song
- Abstract summary: We propose a scalable and diversity-enhanced federated distillation scheme, FedSDD, for federated learning.
FedSDD decouples the training complexity from the number of clients to enhance the scalability, and builds the ensemble from a set of aggregated models.
Experiment results show that FedSDD outperforms other FL methods, including FedAvg and FedDF, on the benchmark datasets.
- Score: 15.39242780506777
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, innovative model aggregation methods based on knowledge
distillation (KD) have been proposed for federated learning (FL). These methods
not only improved the robustness of model aggregation over heterogeneous
learning environment, but also allowed training heterogeneous models on client
devices. However, the scalability of existing methods is not satisfactory,
because the training cost on the server increases with the number of clients,
which limits their application in large scale systems. Furthermore, the
ensemble of existing methods is built from a set of client models initialized
from the same checkpoint, causing low diversity. In this paper, we propose a
scalable and diversity-enhanced federated distillation scheme, FedSDD, which
decouples the training complexity from the number of clients to enhance the
scalability, and builds the ensemble from a set of aggregated models with
enhanced diversity. In particular, the teacher model in FedSDD is an ensemble
built by a small group of aggregated (global) models, instead of all client
models, such that the computation cost will not scale with the number of
clients. Furthermore, to enhance diversity, FedSDD only performs KD to enhance
one of the global models, i.e., the \textit{main global model}, which improves
the performance of both the ensemble and the main global model. While
partitioning client model into more groups allow building an ensemble with more
aggregated models, the convergence of individual aggregated models will be slow
down. We introduce the temporal ensembling which leverage the issues, and
provide significant improvement with the heterogeneous settings. Experiment
results show that FedSDD outperforms other FL methods, including FedAvg and
FedDF, on the benchmark datasets.
Related papers
- Multi-Level Additive Modeling for Structured Non-IID Federated Learning [54.53672323071204]
We train models organized in a multi-level structure, called Multi-level Additive Models (MAM)'', for better knowledge-sharing across heterogeneous clients.
In federated MAM (FeMAM), each client is assigned to at most one model per level and its personalized prediction sums up the outputs of models assigned to it across all levels.
Experiments show that FeMAM surpasses existing clustered FL and personalized FL methods in various non-IID settings.
arXiv Detail & Related papers (2024-05-26T07:54:53Z) - An Aggregation-Free Federated Learning for Tackling Data Heterogeneity [50.44021981013037]
Federated Learning (FL) relies on the effectiveness of utilizing knowledge from distributed datasets.
Traditional FL methods adopt an aggregate-then-adapt framework, where clients update local models based on a global model aggregated by the server from the previous training round.
We introduce FedAF, a novel aggregation-free FL algorithm.
arXiv Detail & Related papers (2024-04-29T05:55:23Z) - DA-PFL: Dynamic Affinity Aggregation for Personalized Federated Learning [13.393529840544117]
Existing personalized federated learning models prefer to aggregate similar clients with similar data distribution to improve the performance of learning models.
We propose a novel Dynamic Affinity-based Personalized Federated Learning model (DA-PFL) to alleviate the class imbalanced problem.
arXiv Detail & Related papers (2024-03-14T11:12:10Z) - Enhancing One-Shot Federated Learning Through Data and Ensemble
Co-Boosting [76.64235084279292]
One-shot Federated Learning (OFL) has become a promising learning paradigm, enabling the training of a global server model via a single communication round.
We introduce a novel framework, Co-Boosting, in which synthesized data and the ensemble model mutually enhance each other progressively.
arXiv Detail & Related papers (2024-02-23T03:15:10Z) - Mitigating System Bias in Resource Constrained Asynchronous Federated
Learning Systems [2.8790600498444032]
We propose a dynamic global model aggregation method within Asynchronous Federated Learning (AFL) deployments.
Our method scores and adjusts the weighting of client model updates based on their upload frequency to accommodate differences in device capabilities.
arXiv Detail & Related papers (2024-01-24T10:51:15Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Closing the Gap between Client and Global Model Performance in
Heterogeneous Federated Learning [2.1044900734651626]
We show how the chosen approach for training custom client models has an impact on the global model.
We propose a new approach that combines KD and Learning without Forgetting (LwoF) to produce improved personalised models.
arXiv Detail & Related papers (2022-11-07T11:12:57Z) - FedDM: Iterative Distribution Matching for Communication-Efficient
Federated Learning [87.08902493524556]
Federated learning(FL) has recently attracted increasing attention from academia and industry.
We propose FedDM to build the global training objective from multiple local surrogate functions.
In detail, we construct synthetic sets of data on each client to locally match the loss landscape from original data.
arXiv Detail & Related papers (2022-07-20T04:55:18Z) - Heterogeneous Ensemble Knowledge Transfer for Training Large Models in
Federated Learning [22.310090483499035]
Federated learning (FL) enables edge-devices to collaboratively learn a model without disclosing their private data to a central aggregating server.
Most existing FL algorithms require models of identical architecture to be deployed across the clients and server.
We propose a novel ensemble knowledge transfer method named Fed-ET in which small models are trained on clients, and used to train a larger model at the server.
arXiv Detail & Related papers (2022-04-27T05:18:32Z) - Ensemble Distillation for Robust Model Fusion in Federated Learning [72.61259487233214]
Federated Learning (FL) is a machine learning setting where many devices collaboratively train a machine learning model.
In most of the current training schemes the central model is refined by averaging the parameters of the server model and the updated parameters from the client side.
We propose ensemble distillation for model fusion, i.e. training the central classifier through unlabeled data on the outputs of the models from the clients.
arXiv Detail & Related papers (2020-06-12T14:49:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.