Gradient tracking and variance reduction for decentralized optimization
and machine learning
- URL: http://arxiv.org/abs/2002.05373v1
- Date: Thu, 13 Feb 2020 07:17:07 GMT
- Title: Gradient tracking and variance reduction for decentralized optimization
and machine learning
- Authors: Ran Xin, Soummya Kar, Usman A. Khan
- Abstract summary: Decentralized methods to solve finite-sum problems are important in many signal processing and machine learning tasks.
We provide a unified algorithmic framework that combines variance-reduction with gradient tracking to achieve robust performance.
- Score: 19.54092620537586
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Decentralized methods to solve finite-sum minimization problems are important
in many signal processing and machine learning tasks where the data is
distributed over a network of nodes and raw data sharing is not permitted due
to privacy and/or resource constraints. In this article, we review
decentralized stochastic first-order methods and provide a unified algorithmic
framework that combines variance-reduction with gradient tracking to achieve
both robust performance and fast convergence. We provide explicit theoretical
guarantees of the corresponding methods when the objective functions are smooth
and strongly-convex, and show their applicability to non-convex problems via
numerical experiments. Throughout the article, we provide intuitive
illustrations of the main technical ideas by casting appropriate tradeoffs and
comparisons among the methods of interest and by highlighting applications to
decentralized training of machine learning models.
Related papers
- Robust Decentralized Learning with Local Updates and Gradient Tracking [16.46727164965154]
We consider decentralized learning as a network of communicating clients or nodes.
We propose a decentralized minimax optimization method that employs two important data: local updates and gradient tracking.
arXiv Detail & Related papers (2024-05-02T03:03:34Z) - Decentralized Learning Strategies for Estimation Error Minimization with Graph Neural Networks [94.2860766709971]
We address the challenge of sampling and remote estimation for autoregressive Markovian processes in a wireless network with statistically-identical agents.
Our goal is to minimize time-average estimation error and/or age of information with decentralized scalable sampling and transmission policies.
arXiv Detail & Related papers (2024-04-04T06:24:11Z) - Can Decentralized Stochastic Minimax Optimization Algorithms Converge
Linearly for Finite-Sum Nonconvex-Nonconcave Problems? [56.62372517641597]
Decentralized minimax optimization has been actively studied in the past few years due to its application in a wide range machine learning.
This paper develops two novel decentralized minimax optimization algorithms for the non-strongly-nonconcave problem.
arXiv Detail & Related papers (2023-04-24T02:19:39Z) - Federated Learning with a Sampling Algorithm under Isoperimetry [9.990687944474738]
Federated learning uses a set of techniques to efficiently distribute the training of a machine learning algorithm across several devices.
We propose a communication-efficient variant of Langevinvin's sample a posteriori.
arXiv Detail & Related papers (2022-06-02T08:19:03Z) - Communication-Efficient Distributionally Robust Decentralized Learning [23.612400109629544]
Decentralized learning algorithms empower interconnected edge devices to share data and computational resources.
We propose a single decentralized loop descent/ascent algorithm (ADGDA) to solve the underlying minimax optimization problem.
arXiv Detail & Related papers (2022-05-31T09:00:37Z) - Distributed Methods with Compressed Communication for Solving
Variational Inequalities, with Theoretical Guarantees [115.08148491584997]
We present the first theoretically grounded distributed methods for solving variational inequalities and saddle point problems using compressed communication: MASHA1 and MASHA2.
New algorithms support bidirectional compressions, and also can be modified for setting with batches and for federated learning with partial participation of clients.
arXiv Detail & Related papers (2021-10-07T10:04:32Z) - Decentralized Personalized Federated Learning for Min-Max Problems [79.61785798152529]
This paper is the first to study PFL for saddle point problems encompassing a broader range of optimization problems.
We propose new algorithms to address this problem and provide a theoretical analysis of the smooth (strongly) convex-(strongly) concave saddle point problems.
Numerical experiments for bilinear problems and neural networks with adversarial noise demonstrate the effectiveness of the proposed methods.
arXiv Detail & Related papers (2021-06-14T10:36:25Z) - Decentralized Deep Learning using Momentum-Accelerated Consensus [15.333413663982874]
We consider the problem of decentralized deep learning where multiple agents collaborate to learn from a distributed dataset.
We propose and analyze a novel decentralized deep learning algorithm where the agents interact over a fixed communication topology.
Our algorithm is based on the heavy-ball acceleration method used in gradient-based protocol.
arXiv Detail & Related papers (2020-10-21T17:39:52Z) - A general framework for decentralized optimization with first-order
methods [11.50057411285458]
Decentralized optimization to minimize a finite sum of functions over a network of nodes has been a significant focus in control and signal processing research.
The emergence of sophisticated computing and large-scale data science needs have led to a resurgence of activity in this area.
We discuss decentralized first-order gradient methods, which have found tremendous success in control, signal processing, and machine learning problems.
arXiv Detail & Related papers (2020-09-12T17:52:10Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z) - Quantized Decentralized Stochastic Learning over Directed Graphs [52.94011236627326]
We consider a decentralized learning problem where data points are distributed among computing nodes communicating over a directed graph.
As the model size gets large, decentralized learning faces a major bottleneck that is the communication load due to each node transmitting messages (model updates) to its neighbors.
We propose the quantized decentralized learning algorithm over directed graphs that is based on the push-sum algorithm in decentralized consensus optimization.
arXiv Detail & Related papers (2020-02-23T18:25:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.