Distributed Stochastic Consensus Optimization with Momentum for
Nonconvex Nonsmooth Problems
- URL: http://arxiv.org/abs/2011.05082v1
- Date: Tue, 10 Nov 2020 13:12:21 GMT
- Title: Distributed Stochastic Consensus Optimization with Momentum for
Nonconvex Nonsmooth Problems
- Authors: Zhiguo Wang, Jiawei Zhang, Tsung-Hui Chang, Jian Li and Zhi-Quan Luo
- Abstract summary: This paper presents a new distributed optimization algorithm for non-smooth problems.
We show that the proposed algorithm can achieve an overcal communication.
Experiments are presented to illustrate the effectiveness of the proposed algorithms.
- Score: 45.88640334610308
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While many distributed optimization algorithms have been proposed for solving
smooth or convex problems over the networks, few of them can handle non-convex
and non-smooth problems. Based on a proximal primal-dual approach, this paper
presents a new (stochastic) distributed algorithm with Nesterov momentum for
accelerated optimization of non-convex and non-smooth problems. Theoretically,
we show that the proposed algorithm can achieve an $\epsilon$-stationary
solution under a constant step size with $\mathcal{O}(1/\epsilon^2)$
computation complexity and $\mathcal{O}(1/\epsilon)$ communication complexity.
When compared to the existing gradient tracking based methods, the proposed
algorithm has the same order of computation complexity but lower order of
communication complexity. To the best of our knowledge, the presented result is
the first stochastic algorithm with the $\mathcal{O}(1/\epsilon)$ communication
complexity for non-convex and non-smooth problems. Numerical experiments for a
distributed non-convex regression problem and a deep neural network based
classification problem are presented to illustrate the effectiveness of the
proposed algorithms.
Related papers
- Accelerated First-Order Optimization under Nonlinear Constraints [73.2273449996098]
We exploit between first-order algorithms for constrained optimization and non-smooth systems to design a new class of accelerated first-order algorithms.
An important property of these algorithms is that constraints are expressed in terms of velocities instead of sparse variables.
arXiv Detail & Related papers (2023-02-01T08:50:48Z) - An Accelerated Stochastic Algorithm for Solving the Optimal Transport
Problem [2.1485350418225244]
A primal-dual accelerated gradient descent with variance reduction algorithm (PDASGD) is proposed to solve linear-constrained optimization problems.
PDASGD enjoys the best-known computational complexity -- $widetildemathcalO(n2/epsilon)$, where $n$ is the number of atoms, and $epsilon>0$ is the accuracy.
arXiv Detail & Related papers (2022-03-02T01:16:10Z) - Lower Bounds and Optimal Algorithms for Smooth and Strongly Convex
Decentralized Optimization Over Time-Varying Networks [79.16773494166644]
We consider the task of minimizing the sum of smooth and strongly convex functions stored in a decentralized manner across the nodes of a communication network.
We design two optimal algorithms that attain these lower bounds.
We corroborate the theoretical efficiency of these algorithms by performing an experimental comparison with existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-08T15:54:44Z) - Randomized Stochastic Variance-Reduced Methods for Stochastic Bilevel
Optimization [62.87181271021217]
We consider non-SBO problems that have many applications in machine learning.
This paper proposes fast randomized algorithms for non-SBO problems.
arXiv Detail & Related papers (2021-05-05T18:28:42Z) - Recent Theoretical Advances in Non-Convex Optimization [56.88981258425256]
Motivated by recent increased interest in analysis of optimization algorithms for non- optimization in deep networks and other problems in data, we give an overview of recent results of theoretical optimization algorithms for non- optimization.
arXiv Detail & Related papers (2020-12-11T08:28:51Z) - Optimal Algorithms for Convex Nested Stochastic Composite Optimization [13.200502573462712]
convex nested composite optimization (NSCO) has received considerable attention for its applications in reinforcement learning and risk-averse optimization.
The current NSCO algorithms have worse oracle complexities, by orders of magnitude, than those for simpler composite optimization problems without the nested structure.
We develop order-optimal algorithms for the convex NSCO problem constructed from an arbitrary composition of smooth, structured non-smooth and general non-smooth layer functions.
arXiv Detail & Related papers (2020-11-19T19:22:58Z) - Single-Timescale Stochastic Nonconvex-Concave Optimization for Smooth
Nonlinear TD Learning [145.54544979467872]
We propose two single-timescale single-loop algorithms that require only one data point each step.
Our results are expressed in a form of simultaneous primal and dual side convergence.
arXiv Detail & Related papers (2020-08-23T20:36:49Z) - Optimal and Practical Algorithms for Smooth and Strongly Convex
Decentralized Optimization [21.555331273873175]
We consider the task of decentralized minimization of the sum of smooth strongly convex functions stored across the nodes of a network.
We propose two new algorithms for this decentralized optimization problem and equip them with complexity guarantees.
arXiv Detail & Related papers (2020-06-21T11:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.