Optimization-Based GenQSGD for Federated Edge Learning
- URL: http://arxiv.org/abs/2110.12987v1
- Date: Mon, 25 Oct 2021 14:25:11 GMT
- Title: Optimization-Based GenQSGD for Federated Edge Learning
- Authors: Yangchen Li, Ying Cui, Vincent Lau
- Abstract summary: We present a generalized parallel mini-batch convergence descent (SGD) algorithm for federated learning (FL)
We optimize the algorithm parameters to minimize the energy cost under the time convergence error.
Results demonstrate the significant gains over existing FL algorithms.
- Score: 12.371264770814097
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Optimal algorithm design for federated learning (FL) remains an open problem.
This paper explores the full potential of FL in practical edge computing
systems where workers may have different computation and communication
capabilities, and quantized intermediate model updates are sent between the
server and workers. First, we present a general quantized parallel mini-batch
stochastic gradient descent (SGD) algorithm for FL, namely GenQSGD, which is
parameterized by the number of global iterations, the numbers of local
iterations at all workers, and the mini-batch size. We also analyze its
convergence error for any choice of the algorithm parameters. Then, we optimize
the algorithm parameters to minimize the energy cost under the time constraint
and convergence error constraint. The optimization problem is a challenging
non-convex problem with non-differentiable constraint functions. We propose an
iterative algorithm to obtain a KKT point using advanced optimization
techniques. Numerical results demonstrate the significant gains of GenQSGD over
existing FL algorithms and reveal the importance of optimally designing FL
algorithms.
Related papers
- Federated Conditional Stochastic Optimization [110.513884892319]
Conditional optimization has found in a wide range of machine learning tasks, such as in-variant learning tasks, AUPRC, andAML.
This paper proposes algorithms for distributed federated learning.
arXiv Detail & Related papers (2023-10-04T01:47:37Z) - GQFedWAvg: Optimization-Based Quantized Federated Learning in General
Edge Computing Systems [11.177402054314674]
The optimal implementation of federated learning (FL) in practical edge computing has been an outstanding problem.
We propose an optimization quantized FL algorithm, which can appropriately fit a general edge computing system with uniform or nonuniform computing and communication systems.
arXiv Detail & Related papers (2023-06-13T02:18:24Z) - Accelerated First-Order Optimization under Nonlinear Constraints [73.2273449996098]
We exploit between first-order algorithms for constrained optimization and non-smooth systems to design a new class of accelerated first-order algorithms.
An important property of these algorithms is that constraints are expressed in terms of velocities instead of sparse variables.
arXiv Detail & Related papers (2023-02-01T08:50:48Z) - Adaptive Federated Minimax Optimization with Lower Complexities [82.51223883622552]
We propose an efficient adaptive minimax optimization algorithm (i.e., AdaFGDA) to solve these minimax problems.
It builds our momentum-based reduced and localSGD techniques, and it flexibly incorporate various adaptive learning rates.
arXiv Detail & Related papers (2022-11-14T12:32:18Z) - Optimization-based Block Coordinate Gradient Coding for Mitigating
Partial Stragglers in Distributed Learning [58.91954425047425]
This paper aims to design a new gradient coding scheme for mitigating partial stragglers in distributed learning.
We propose a gradient coordinate coding scheme with L coding parameters representing L possibly different diversities for the L coordinates, which generates most gradient coding schemes.
arXiv Detail & Related papers (2022-06-06T09:25:40Z) - An Optimization Framework for Federated Edge Learning [11.007444733506714]
This paper considers edge computing system where server and workers have possibly different computing and communication capabilities.
We first present a general FL algorithm, namely GenQSGD, parameterized by the numbers of global and local iterations, mini-batch size, and step size sequence.
arXiv Detail & Related papers (2021-11-26T14:47:32Z) - DESTRESS: Computation-Optimal and Communication-Efficient Decentralized
Nonconvex Finite-Sum Optimization [43.31016937305845]
Internet-of-things, networked sensing, autonomous systems and federated learning call for decentralized algorithms for finite-sum optimizations.
We develop DEcentralized STochastic REcurSive methodDESTRESS for non finite-sum optimization.
Detailed theoretical and numerical comparisons show that DESTRESS improves upon prior decentralized algorithms.
arXiv Detail & Related papers (2021-10-04T03:17:41Z) - Continuation Newton methods with deflation techniques for global
optimization problems [3.705839280172101]
A global minimum point of an optimization problem is of interest in engineering.
In this article, we consider a new memetic algorithm for this nonlinear largescale problem.
According to our numerical experiments, new algorithm works well for unconstrained unconstrained problems.
arXiv Detail & Related papers (2021-07-29T09:53:49Z) - Lower Bounds and Optimal Algorithms for Smooth and Strongly Convex
Decentralized Optimization Over Time-Varying Networks [79.16773494166644]
We consider the task of minimizing the sum of smooth and strongly convex functions stored in a decentralized manner across the nodes of a communication network.
We design two optimal algorithms that attain these lower bounds.
We corroborate the theoretical efficiency of these algorithms by performing an experimental comparison with existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-08T15:54:44Z) - Decentralized and Model-Free Federated Learning: Consensus-Based
Distillation in Function Space [7.627597166844701]
This paper proposes a decentralized FL scheme for IoE devices connected via multi-hop networks.
It shows that CMFD achieves higher stability than parameter aggregation methods.
arXiv Detail & Related papers (2021-04-01T09:17:20Z) - Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization [71.03797261151605]
Adaptivity is an important yet under-studied property in modern optimization theory.
Our algorithm is proved to achieve the best-available convergence for non-PL objectives simultaneously while outperforming existing algorithms for PL objectives.
arXiv Detail & Related papers (2020-02-13T05:42:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.