Proof of Swarm Based Ensemble Learning for Federated Learning
Applications
- URL: http://arxiv.org/abs/2212.14050v1
- Date: Wed, 28 Dec 2022 13:53:34 GMT
- Title: Proof of Swarm Based Ensemble Learning for Federated Learning
Applications
- Authors: Ali Raza, Kim Phuc Tran, Ludovic Koehl, Shujun Li
- Abstract summary: In federated learning it is not feasible to apply centralised ensemble learning directly due to privacy concerns.
Most distributed consensus algorithms, such as Byzantine fault tolerance (BFT), do not normally perform well in such applications.
We propose PoSw, a novel distributed consensus algorithm for ensemble learning in a federated setting.
- Score: 3.2536767864585663
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Ensemble learning combines results from multiple machine learning models in
order to provide a better and optimised predictive model with reduced bias,
variance and improved predictions. However, in federated learning it is not
feasible to apply centralised ensemble learning directly due to privacy
concerns. Hence, a mechanism is required to combine results of local models to
produce a global model. Most distributed consensus algorithms, such as
Byzantine fault tolerance (BFT), do not normally perform well in such
applications. This is because, in such methods predictions of some of the peers
are disregarded, so a majority of peers can win without even considering other
peers' decisions. Additionally, the confidence score of the result of each peer
is not normally taken into account, although it is an important feature to
consider for ensemble learning. Moreover, the problem of a tie event is often
left un-addressed by methods such as BFT. To fill these research gaps, we
propose PoSw (Proof of Swarm), a novel distributed consensus algorithm for
ensemble learning in a federated setting, which was inspired by particle swarm
based algorithms for solving optimisation problems. The proposed algorithm is
theoretically proved to always converge in a relatively small number of steps
and has mechanisms to resolve tie events while trying to achieve sub-optimum
solutions. We experimentally validated the performance of the proposed
algorithm using ECG classification as an example application in healthcare,
showing that the ensemble learning model outperformed all local models and even
the FL-based global model. To the best of our knowledge, the proposed algorithm
is the first attempt to make consensus over the output results of distributed
models trained using federated learning.
Related papers
- EnsIR: An Ensemble Algorithm for Image Restoration via Gaussian Mixture Models [70.60381055741391]
Image restoration challenges related to illposed problems, resulting in deviations between single model predictions and ground-truths.
Ensemble learning aims to address these deviations by combining the predictions of multiple base models.
We employ an expectation (EM)-based algorithm to estimate ensemble weights for prediction candidates.
Our algorithm is model-agnostic and training-free, allowing seamless integration and enhancement of various pre-trained image restoration models.
arXiv Detail & Related papers (2024-10-30T12:16:35Z) - A Kernel Perspective on Distillation-based Collaborative Learning [8.971234046933349]
We propose a nonparametric collaborative learning algorithm that does not directly share local data or models in statistically heterogeneous environments.
Inspired by our theoretical results, we also propose a practical distillation-based collaborative learning algorithm based on neural network architecture.
arXiv Detail & Related papers (2024-10-23T06:40:13Z) - LoRA-Ensemble: Efficient Uncertainty Modelling for Self-attention Networks [52.46420522934253]
We introduce LoRA-Ensemble, a parameter-efficient deep ensemble method for self-attention networks.
By employing a single pre-trained self-attention network with weights shared across all members, we train member-specific low-rank matrices for the attention projections.
Our method exhibits superior calibration compared to explicit ensembles and achieves similar or better accuracy across various prediction tasks and datasets.
arXiv Detail & Related papers (2024-05-23T11:10:32Z) - Vanishing Variance Problem in Fully Decentralized Neural-Network Systems [0.8212195887472242]
Federated learning and gossip learning are emerging methodologies designed to mitigate data privacy concerns.
Our research introduces a variance-corrected model averaging algorithm.
Our simulation results demonstrate that our approach enables gossip learning to achieve convergence efficiency comparable to that of federated learning.
arXiv Detail & Related papers (2024-04-06T12:49:20Z) - FedCBO: Reaching Group Consensus in Clustered Federated Learning through
Consensus-based Optimization [1.911678487931003]
Federated learning seeks to integrate the training learning models from multiple users, each user having their own data set, in a way that is sensitive to data privacy and to communication loss constraints.
In this paper, we propose a novel solution to a global, clustered problem of federated learning that is inspired by ideas in consensus-based optimization (CBO)
Our new CBO-type method is based on a system of interacting particles that is oblivious to group.
arXiv Detail & Related papers (2023-05-04T15:02:09Z) - Joint Training of Deep Ensembles Fails Due to Learner Collusion [61.557412796012535]
Ensembles of machine learning models have been well established as a powerful method of improving performance over a single model.
Traditionally, ensembling algorithms train their base learners independently or sequentially with the goal of optimizing their joint performance.
We show that directly minimizing the loss of the ensemble appears to rarely be applied in practice.
arXiv Detail & Related papers (2023-01-26T18:58:07Z) - Deep Negative Correlation Classification [82.45045814842595]
Existing deep ensemble methods naively train many different models and then aggregate their predictions.
We propose deep negative correlation classification (DNCC)
DNCC yields a deep classification ensemble where the individual estimator is both accurate and negatively correlated.
arXiv Detail & Related papers (2022-12-14T07:35:20Z) - Faster Adaptive Federated Learning [84.38913517122619]
Federated learning has attracted increasing attention with the emergence of distributed data.
In this paper, we propose an efficient adaptive algorithm (i.e., FAFED) based on momentum-based variance reduced technique in cross-silo FL.
arXiv Detail & Related papers (2022-12-02T05:07:50Z) - Federated Learning Aggregation: New Robust Algorithms with Guarantees [63.96013144017572]
Federated learning has been recently proposed for distributed model training at the edge.
This paper presents a complete general mathematical convergence analysis to evaluate aggregation strategies in a federated learning framework.
We derive novel aggregation algorithms which are able to modify their model architecture by differentiating client contributions according to the value of their losses.
arXiv Detail & Related papers (2022-05-22T16:37:53Z) - Towards Model Agnostic Federated Learning Using Knowledge Distillation [9.947968358822951]
In this work, we initiate a theoretical study of model agnostic communication protocols.
We focus on the setting where the two agents are attempting to perform kernel regression using different kernels.
Our study yields a surprising result -- the most natural algorithm of using alternating knowledge distillation (AKD) imposes overly strong regularization.
arXiv Detail & Related papers (2021-10-28T15:27:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.