Reducing Impacts of System Heterogeneity in Federated Learning using
Weight Update Magnitudes
- URL: http://arxiv.org/abs/2208.14808v1
- Date: Tue, 30 Aug 2022 00:39:06 GMT
- Title: Reducing Impacts of System Heterogeneity in Federated Learning using
Weight Update Magnitudes
- Authors: Irene Wang
- Abstract summary: Federated learning enables machine learning models to train locally on each handheld device while only synchronizing their neuron updates with a server.
This results in the training time of federated learning tasks being dictated by a few low-performance straggler devices.
In this work, we aim to mitigate the performance bottleneck of federated learning by dynamically forming sub-models for stragglers based on their performance and accuracy feedback.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The widespread adoption of handheld devices have fueled rapid growth in new
applications. Several of these new applications employ machine learning models
to train on user data that is typically private and sensitive. Federated
Learning enables machine learning models to train locally on each handheld
device while only synchronizing their neuron updates with a server. While this
enables user privacy, technology scaling and software advancements have
resulted in handheld devices with varying performance capabilities. This
results in the training time of federated learning tasks to be dictated by a
few low-performance straggler devices, essentially becoming a bottleneck to the
entire training process. In this work, we aim to mitigate the performance
bottleneck of federated learning by dynamically forming sub-models for
stragglers based on their performance and accuracy feedback. To this end, we
offer the Invariant Dropout, a dynamic technique that forms a sub-model based
on the neuron update threshold. Invariant Dropout uses neuron updates from the
non-straggler clients to develop a tailored sub-models for each straggler
during each training iteration. All corresponding weights which have a
magnitude less than the threshold are dropped for the iteration. We evaluate
Invariant Dropout using five real-world mobile clients. Our evaluations show
that Invariant Dropout obtains a maximum accuracy gain of 1.4% points over
state-of-the-art Ordered Dropout while mitigating performance bottlenecks of
stragglers.
Related papers
- Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters [65.15700861265432]
We present a parameter-efficient continual learning framework to alleviate long-term forgetting in incremental learning with vision-language models.
Our approach involves the dynamic expansion of a pre-trained CLIP model, through the integration of Mixture-of-Experts (MoE) adapters.
To preserve the zero-shot recognition capability of vision-language models, we introduce a Distribution Discriminative Auto-Selector.
arXiv Detail & Related papers (2024-03-18T08:00:23Z) - FedSPU: Personalized Federated Learning for Resource-constrained Devices with Stochastic Parameter Update [0.27309692684728615]
Federated Dropout has emerged as a popular strategy to address this challenge.
We propose federated learning with parameter update (FedSPU)
Experimental results demonstrate that FedSPU outperforms federated dropout by 7.57% on average in terms of accuracy.
arXiv Detail & Related papers (2024-03-18T04:31:38Z) - Efficient Asynchronous Federated Learning with Sparsification and
Quantization [55.6801207905772]
Federated Learning (FL) is attracting more and more attention to collaboratively train a machine learning model without transferring raw data.
FL generally exploits a parameter server and a large number of edge devices during the whole process of the model training.
We propose TEASQ-Fed to exploit edge devices to asynchronously participate in the training process by actively applying for tasks.
arXiv Detail & Related papers (2023-12-23T07:47:07Z) - FedYolo: Augmenting Federated Learning with Pretrained Transformers [61.56476056444933]
In this work, we investigate pretrained transformers (PTF) to achieve on-device learning goals.
We show that larger scale shrinks the accuracy gaps between alternative approaches and improves robustness.
Finally, it enables clients to solve multiple unrelated tasks simultaneously using a single PTF.
arXiv Detail & Related papers (2023-07-10T21:08:52Z) - FLuID: Mitigating Stragglers in Federated Learning using Invariant
Dropout [1.8262547855491458]
Federated Learning allows machine learning models to train locally on individual mobile devices, synchronizing model updates via a shared server.
As a result, straggler devices with lower performance often dictate the overall training time in FL.
We introduce Invariant Dropout, a method that extracts a sub-model based on the weight update threshold.
We develop an adaptive training framework, Federated Learning using Invariant Dropout.
arXiv Detail & Related papers (2023-07-05T19:53:38Z) - Exploiting Features and Logits in Heterogeneous Federated Learning [0.2538209532048866]
Federated learning (FL) facilitates the management of edge devices to collaboratively train a shared model.
We propose a novel data-free FL method that supports heterogeneous client models by managing features and logits.
Unlike Felo, the server has a conditional VAE in Velo, which is used for training mid-level features and generating synthetic features according to the labels.
arXiv Detail & Related papers (2022-10-27T15:11:46Z) - Applied Federated Learning: Architectural Design for Robust and
Efficient Learning in Privacy Aware Settings [0.8454446648908585]
The classical machine learning paradigm requires the aggregation of user data in a central location.
Centralization of data poses risks, including a heightened risk of internal and external security incidents.
Federated learning with differential privacy is designed to avoid the server-side centralization pitfall.
arXiv Detail & Related papers (2022-06-02T00:30:04Z) - Acceleration of Federated Learning with Alleviated Forgetting in Local
Training [61.231021417674235]
Federated learning (FL) enables distributed optimization of machine learning models while protecting privacy.
We propose FedReg, an algorithm to accelerate FL with alleviated knowledge forgetting in the local training stage.
Our experiments demonstrate that FedReg not only significantly improves the convergence rate of FL, especially when the neural network architecture is deep.
arXiv Detail & Related papers (2022-03-05T02:31:32Z) - Fast-Convergent Federated Learning [82.32029953209542]
Federated learning is a promising solution for distributing machine learning tasks through modern networks of mobile devices.
We propose a fast-convergent federated learning algorithm, called FOLB, which performs intelligent sampling of devices in each round of model training.
arXiv Detail & Related papers (2020-07-26T14:37:51Z) - UVeQFed: Universal Vector Quantization for Federated Learning [179.06583469293386]
Federated learning (FL) is an emerging approach to train such learning models without requiring the users to share their possibly private labeled data.
In FL, each user trains its copy of the learning model locally. The server then collects the individual updates and aggregates them into a global model.
We show that combining universal vector quantization methods with FL yields a decentralized training system in which the compression of the trained models induces only a minimum distortion.
arXiv Detail & Related papers (2020-06-05T07:10:22Z) - MDLdroid: a ChainSGD-reduce Approach to Mobile Deep Learning for
Personal Mobile Sensing [14.574274428615666]
Running deep learning on devices offers several advantages including data privacy preservation and low-latency response for both model robustness and update.
Personal mobile sensing applications are mostly user-specific and highly affected by environment.
We present MDLdroid, a novel decentralized mobile deep learning framework to enable resource-aware on-device collaborative learning.
arXiv Detail & Related papers (2020-02-07T16:55:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.