Optimal Epoch Stochastic Gradient Descent Ascent Methods for Min-Max
Optimization
- URL: http://arxiv.org/abs/2002.05309v2
- Date: Wed, 17 Jun 2020 09:02:49 GMT
- Title: Optimal Epoch Stochastic Gradient Descent Ascent Methods for Min-Max
Optimization
- Authors: Yan Yan and Yi Xu and Qihang Lin and Wei Liu and Tianbao Yang
- Abstract summary: Epoch gradient descent method (a.k.a. Epoch-GD) proposed by Hazan and Kale (2011) was deemed a breakthrough for strongly convex minimization.
Our result is the first one that shows Epoch-GDA can achieve the optimal rate of $O (1/T)$ for the duality gap of SCSC min-max problems.
- Score: 61.66927778694731
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Epoch gradient descent method (a.k.a. Epoch-GD) proposed by Hazan and Kale
(2011) was deemed a breakthrough for stochastic strongly convex minimization,
which achieves the optimal convergence rate of $O(1/T)$ with $T$ iterative
updates for the {\it objective gap}. However, its extension to solving
stochastic min-max problems with strong convexity and strong concavity still
remains open, and it is still unclear whether a fast rate of $O(1/T)$ for the
{\it duality gap} is achievable for stochastic min-max optimization under
strong convexity and strong concavity. Although some recent studies have
proposed stochastic algorithms with fast convergence rates for min-max
problems, they require additional assumptions about the problem, e.g.,
smoothness, bi-linear structure, etc. In this paper, we bridge this gap by
providing a sharp analysis of epoch-wise stochastic gradient descent ascent
method (referred to as Epoch-GDA) for solving strongly convex strongly concave
(SCSC) min-max problems, without imposing any additional assumption about
smoothness or the function's structure. To the best of our knowledge, our
result is the first one that shows Epoch-GDA can achieve the optimal rate of
$O(1/T)$ for the duality gap of general SCSC min-max problems. We emphasize
that such generalization of Epoch-GD for strongly convex minimization problems
to Epoch-GDA for SCSC min-max problems is non-trivial and requires novel
technical analysis. Moreover, we notice that the key lemma can also be used for
proving the convergence of Epoch-GDA for weakly-convex strongly-concave min-max
problems, leading to a nearly optimal complexity without resorting to
smoothness or other structural conditions.
Related papers
- Methods for Convex $(L_0,L_1)$-Smooth Optimization: Clipping, Acceleration, and Adaptivity [50.25258834153574]
We focus on the class of (strongly) convex $(L0)$-smooth functions and derive new convergence guarantees for several existing methods.
In particular, we derive improved convergence rates for Gradient Descent with smoothnessed Gradient Clipping and for Gradient Descent with Polyak Stepsizes.
arXiv Detail & Related papers (2024-09-23T13:11:37Z) - Projective Proximal Gradient Descent for A Class of Nonconvex Nonsmooth Optimization Problems: Fast Convergence Without Kurdyka-Lojasiewicz (KL) Property [19.988762532185884]
Non and nonsmooth optimization problems are important and challenging for learning.
In this paper, we show a new analysis showing fast convergence of PPGD.
arXiv Detail & Related papers (2023-04-20T17:39:24Z) - Stochastic Nonsmooth Convex Optimization with Heavy-Tailed Noises:
High-Probability Bound, In-Expectation Rate and Initial Distance Adaptation [22.758674468435302]
In a heavy-tailed noise regime, the difference between the gradient and the true rate is assumed to have a finite $p-th moment.
This paper provides a comprehensive analysis of nonsmooth convex optimization with heavy-tailed noises.
arXiv Detail & Related papers (2023-03-22T03:05:28Z) - Near-Optimal Algorithms for Making the Gradient Small in Stochastic
Minimax Optimization [14.719077076351377]
We study the problem of finding a near-stationary point for smooth minimax optimization.
We show that a novel algorithm called Recursive IteratioNRAIN achieves both convex-concave and strongly-concave cases.
arXiv Detail & Related papers (2022-08-11T16:55:26Z) - Optimal Extragradient-Based Bilinearly-Coupled Saddle-Point Optimization [116.89941263390769]
We consider the smooth convex-concave bilinearly-coupled saddle-point problem, $min_mathbfxmax_mathbfyF(mathbfx) + H(mathbfx,mathbfy)$, where one has access to first-order oracles for $F$, $G$ as well as the bilinear coupling function $H$.
We present a emphaccelerated gradient-extragradient (AG-EG) descent-ascent algorithm that combines extragrad
arXiv Detail & Related papers (2022-06-17T06:10:20Z) - Convergence Rates of Two-Time-Scale Gradient Descent-Ascent Dynamics for
Solving Nonconvex Min-Max Problems [2.0305676256390934]
We characterize the finite-time performance of the continuous-time variant of simultaneous gradient descent-ascent algorithm.
Our results on the behavior of continuous-time algorithm may be used to enhance the convergence properties of its discrete-time counterpart.
arXiv Detail & Related papers (2021-12-17T15:51:04Z) - Faster Algorithm and Sharper Analysis for Constrained Markov Decision
Process [56.55075925645864]
The problem of constrained decision process (CMDP) is investigated, where an agent aims to maximize the expected accumulated discounted reward subject to multiple constraints.
A new utilities-dual convex approach is proposed with novel integration of three ingredients: regularized policy, dual regularizer, and Nesterov's gradient descent dual.
This is the first demonstration that nonconcave CMDP problems can attain the lower bound of $mathcal O (1/epsilon)$ for all complexity optimization subject to convex constraints.
arXiv Detail & Related papers (2021-10-20T02:57:21Z) - A Momentum-Assisted Single-Timescale Stochastic Approximation Algorithm
for Bilevel Optimization [112.59170319105971]
We propose a new algorithm -- the Momentum- Single-timescale Approximation (MSTSA) -- for tackling problems.
MSTSA allows us to control the error in iterations due to inaccurate solution to the lower level subproblem.
arXiv Detail & Related papers (2021-02-15T07:10:33Z) - Fast Objective & Duality Gap Convergence for Non-Convex Strongly-Concave
Min-Max Problems with PL Condition [52.08417569774822]
This paper focuses on methods for solving smooth non-concave min-max problems, which have received increasing attention due to deep learning (e.g., deep AUC)
arXiv Detail & Related papers (2020-06-12T00:32:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.