SWIFT: Rapid Decentralized Federated Learning via Wait-Free Model
Communication
- URL: http://arxiv.org/abs/2210.14026v1
- Date: Tue, 25 Oct 2022 14:01:21 GMT
- Title: SWIFT: Rapid Decentralized Federated Learning via Wait-Free Model
Communication
- Authors: Marco Bornstein, Tahseen Rabbani, Evan Wang, Amrit Singh Bedi, and
Furong Huang
- Abstract summary: We show that SWIFT converges faster with respect to run-time due to its wait-free structure.
SWIFT produces loss levels for image classification, over IID and non-IID data settings, upwards 50% faster than existing SOTA algorithms.
- Score: 11.763368822546468
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The decentralized Federated Learning (FL) setting avoids the role of a
potentially unreliable or untrustworthy central host by utilizing groups of
clients to collaboratively train a model via localized training and
model/gradient sharing. Most existing decentralized FL algorithms require
synchronization of client models where the speed of synchronization depends
upon the slowest client. In this work, we propose SWIFT: a novel wait-free
decentralized FL algorithm that allows clients to conduct training at their own
speed. Theoretically, we prove that SWIFT matches the gold-standard iteration
convergence rate $\mathcal{O}(1/\sqrt{T})$ of parallel stochastic gradient
descent for convex and non-convex smooth optimization (total iterations $T$).
Furthermore, we provide theoretical results for IID and non-IID settings
without any bounded-delay assumption for slow clients which is required by
other asynchronous decentralized FL algorithms. Although SWIFT achieves the
same iteration convergence rate with respect to $T$ as other state-of-the-art
(SOTA) parallel stochastic algorithms, it converges faster with respect to
run-time due to its wait-free structure. Our experimental results demonstrate
that SWIFT's run-time is reduced due to a large reduction in communication time
per epoch, which falls by an order of magnitude compared to synchronous
counterparts. Furthermore, SWIFT produces loss levels for image classification,
over IID and non-IID data settings, upwards of 50% faster than existing SOTA
algorithms.
Related papers
- Asynchronous Byzantine Federated Learning [4.6792910030704515]
Federated learning (FL) enables a set of geographically distributed clients to collectively train a model through a server.
Our solution is one of the first Byzantine-resilient and asynchronous FL algorithms.
We compare the performance of our solution with state-of-the-art algorithms on both image and text datasets.
arXiv Detail & Related papers (2024-06-03T15:29:38Z) - FedAST: Federated Asynchronous Simultaneous Training [27.492821176616815]
Federated Learning (FL) enables devices or clients to collaboratively train machine learning (ML) models without sharing their private data.
Much of the existing work in FL focuses on efficiently learning a model for a single task.
In this paper, we propose simultaneous training of multiple FL models using a common set of datasets.
arXiv Detail & Related papers (2024-06-01T05:14:20Z) - Robust Model Aggregation for Heterogeneous Federated Learning: Analysis and Optimizations [35.58487905412915]
We propose a time-driven SFL (T-SFL) framework for heterogeneous systems.
To evaluate the learning performance of T-SFL, we provide an upper bound on the global loss function.
We develop a discriminative model selection algorithm that removes local models from clients whose number of iterations falls below a predetermined threshold.
arXiv Detail & Related papers (2024-05-11T11:55:26Z) - Stragglers-Aware Low-Latency Synchronous Federated Learning via Layer-Wise Model Updates [71.81037644563217]
Synchronous federated learning (FL) is a popular paradigm for collaborative edge learning.
As some of the devices may have limited computational resources and varying availability, FL latency is highly sensitive to stragglers.
We propose straggler-aware layer-wise federated learning (SALF) that leverages the optimization procedure of NNs via backpropagation to update the global model in a layer-wise fashion.
arXiv Detail & Related papers (2024-03-27T09:14:36Z) - AEDFL: Efficient Asynchronous Decentralized Federated Learning with
Heterogeneous Devices [61.66943750584406]
We propose an Asynchronous Efficient Decentralized FL framework, i.e., AEDFL, in heterogeneous environments.
First, we propose an asynchronous FL system model with an efficient model aggregation method for improving the FL convergence.
Second, we propose a dynamic staleness-aware model update approach to achieve superior accuracy.
Third, we propose an adaptive sparse training method to reduce communication and computation costs without significant accuracy degradation.
arXiv Detail & Related papers (2023-12-18T05:18:17Z) - Adaptive Federated Pruning in Hierarchical Wireless Networks [69.6417645730093]
Federated Learning (FL) is a privacy-preserving distributed learning framework where a server aggregates models updated by multiple devices without accessing their private datasets.
In this paper, we introduce model pruning for HFL in wireless networks to reduce the neural network scale.
We show that our proposed HFL with model pruning achieves similar learning accuracy compared with the HFL without model pruning and reduces about 50 percent communication cost.
arXiv Detail & Related papers (2023-05-15T22:04:49Z) - Efficient and Light-Weight Federated Learning via Asynchronous
Distributed Dropout [22.584080337157168]
Asynchronous learning protocols have regained attention lately, especially in the Federated Learning (FL) setup.
We propose textttAsyncDrop, a novel asynchronous FL framework that utilizes dropout regularization to handle device heterogeneity in distributed settings.
Overall, textttAsyncDrop achieves better performance compared to state of the art asynchronous methodologies.
arXiv Detail & Related papers (2022-10-28T13:00:29Z) - Semi-Synchronous Personalized Federated Learning over Mobile Edge
Networks [88.50555581186799]
We propose a semi-synchronous PFL algorithm, termed as Semi-Synchronous Personalized FederatedAveraging (PerFedS$2$), over mobile edge networks.
We derive an upper bound of the convergence rate of PerFedS2 in terms of the number of participants per global round and the number of rounds.
Experimental results verify the effectiveness of PerFedS2 in saving training time as well as guaranteeing the convergence of training loss.
arXiv Detail & Related papers (2022-09-27T02:12:43Z) - Time-triggered Federated Learning over Wireless Networks [48.389824560183776]
We present a time-triggered FL algorithm (TT-Fed) over wireless networks.
Our proposed TT-Fed algorithm improves the converged test accuracy by up to 12.5% and 5%, respectively.
arXiv Detail & Related papers (2022-04-26T16:37:29Z) - Stragglers Are Not Disaster: A Hybrid Federated Learning Algorithm with
Delayed Gradients [21.63719641718363]
Federated learning (FL) is a new machine learning framework which trains a joint model across a large amount of decentralized computing devices.
This paper presents a novel FL algorithm, namely Hybrid Federated Learning (HFL), to achieve a learning balance in efficiency and effectiveness.
arXiv Detail & Related papers (2021-02-12T02:27:44Z) - Faster Non-Convex Federated Learning via Global and Local Momentum [57.52663209739171]
textttFedGLOMO is the first (first-order) FLtexttFedGLOMO algorithm.
Our algorithm is provably optimal even with communication between the clients and the server.
arXiv Detail & Related papers (2020-12-07T21:05:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.