Accelerated Zeroth-Order and First-Order Momentum Methods from Mini to
Minimax Optimization
- URL: http://arxiv.org/abs/2008.08170v7
- Date: Mon, 17 Jan 2022 01:35:44 GMT
- Title: Accelerated Zeroth-Order and First-Order Momentum Methods from Mini to
Minimax Optimization
- Authors: Feihu Huang, Shangqian Gao, Jian Pei, Heng Huang
- Abstract summary: We propose a new accelerated zeroth-order momentum (AccZOM) method for black-box mini-optimization where only function values can be obtained.
Meanwhile, we propose an accelerated zeroth-order momentum descent (Acc-MDA) method for black-box minimax optimization, where only function values can be obtained.
In particular, our Acc-MDA can obtain a lower gradient complexity of $tildeO(kappa_y2.5epsilon-3)$ with a batch size $O(kappa_y4)$.
- Score: 133.53164856723782
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In the paper, we propose a class of accelerated zeroth-order and first-order
momentum methods for both nonconvex mini-optimization and minimax-optimization.
Specifically, we propose a new accelerated zeroth-order momentum (Acc-ZOM)
method for black-box mini-optimization where only function values can be
obtained. Moreover, we prove that our Acc-ZOM method achieves a lower query
complexity of $\tilde{O}(d^{3/4}\epsilon^{-3})$ for finding an
$\epsilon$-stationary point, which improves the best known result by a factor
of $O(d^{1/4})$ where $d$ denotes the variable dimension. In particular, our
Acc-ZOM does not need large batches required in the existing zeroth-order
stochastic algorithms. Meanwhile, we propose an accelerated zeroth-order
momentum descent ascent (Acc-ZOMDA) method for black-box minimax optimization,
where only function values can be obtained. Our Acc-ZOMDA obtains a low query
complexity of $\tilde{O}((d_1+d_2)^{3/4}\kappa_y^{4.5}\epsilon^{-3})$ without
requiring large batches for finding an $\epsilon$-stationary point, where $d_1$
and $d_2$ denote variable dimensions and $\kappa_y$ is condition number.
Moreover, we propose an accelerated first-order momentum descent ascent
(Acc-MDA) method for minimax optimization, whose explicit gradients are
accessible. Our Acc-MDA achieves a low gradient complexity of
$\tilde{O}(\kappa_y^{4.5}\epsilon^{-3})$ without requiring large batches for
finding an $\epsilon$-stationary point. In particular, our Acc-MDA can obtain a
lower gradient complexity of $\tilde{O}(\kappa_y^{2.5}\epsilon^{-3})$ with a
batch size $O(\kappa_y^4)$, which improves the best known result by a factor of
$O(\kappa_y^{1/2})$. Extensive experimental results on black-box adversarial
attack to deep neural networks and poisoning attack to logistic regression
demonstrate efficiency of our algorithms.
Related papers
- An Efficient Stochastic Algorithm for Decentralized Nonconvex-Strongly-Concave Minimax Optimization [25.00475462213752]
Decentralized Recursive desc. Method (DREAM)
Concretely, it requires $mathcalO(minminappaappa3eps-3,kappa2 N)$ first-order oracle (SFO) calls and $tildemathcalO(kappa2 epsilon-2) communication rounds.
Our numerical experiments validate the superiority of previous methods.
arXiv Detail & Related papers (2022-12-05T16:09:39Z) - Faster Perturbed Stochastic Gradient Methods for Finding Local Minima [92.99933928528797]
We propose tttPullback, a faster perturbed gradient framework for finding local minima.
We show that Pullback with gradient estimators such as SARAH/SP and STORM can find $(epsilon, epsilon_H)$approximate local minima within $tilde O(epsilon-3 + H-6)$.
The core idea of our framework is a step-size pullback'' scheme to control the average movement of the gradient evaluations.
arXiv Detail & Related papers (2021-10-25T07:20:05Z) - AdaGDA: Faster Adaptive Gradient Descent Ascent Methods for Minimax
Optimization [104.96004056928474]
We propose a class of faster adaptive gradient descent methods for non-strongly-concave minimax problems.
We show that our method reaches a lower sample complexity of $O(kappa2.5epsilon-3)$ with the mini-batch size $O(kappa)$.
arXiv Detail & Related papers (2021-06-30T14:47:09Z) - On Riemannian Gradient-Based Methods for Minimax Problems [24.199289678553896]
We propose a class of Riemanian-based methods to solve minimax problems.
We prove that our RGDA has a sample complexity of $tildeO(kappa4eps-3)$.
We also show that our Acc-RSGDA achieves a lower sample complexity on $tildeO(kappa4eps-3)$.
arXiv Detail & Related papers (2020-10-13T00:54:00Z) - Accelerated Stochastic Gradient-free and Projection-free Methods [37.15461647823691]
We propose an accelerated zeroth-order Frank-Wolfe (Acc-SZOFW) based on a new reduced variance technique of STORM.
To relax the large batches required in the Acc-SZOFW, we further propose a novel accelerated zeroth-order Frank-Wolfe (Acc-SZOFW*) based on a new reduced variance technique of STORM.
arXiv Detail & Related papers (2020-07-16T20:50:15Z) - Streaming Complexity of SVMs [110.63976030971106]
We study the space complexity of solving the bias-regularized SVM problem in the streaming model.
We show that for both problems, for dimensions of $frac1lambdaepsilon$, one can obtain streaming algorithms with spacely smaller than $frac1lambdaepsilon$.
arXiv Detail & Related papers (2020-07-07T17:10:00Z) - Gradient Free Minimax Optimization: Variance Reduction and Faster
Convergence [120.9336529957224]
In this paper, we denote the non-strongly setting on the magnitude of a gradient-free minimax optimization problem.
We show that a novel zeroth-order variance reduced descent algorithm achieves the best known query complexity.
arXiv Detail & Related papers (2020-06-16T17:55:46Z) - Nonconvex Zeroth-Order Stochastic ADMM Methods with Lower Function Query
Complexity [109.54166127479093]
Zeroth-order (a.k.a, derivative-free) methods are a class of effective optimization methods for solving machine learning problems.
In this paper, we propose a class faster faster zerothorder alternating gradient method multipliers (MMADMM) to solve the non finitesum problems.
We show that ZOMMAD methods can achieve a lower function $O(frac13nfrac1)$ for finding an $epsilon$-stationary point.
At the same time, we propose a class faster zerothorder online ADM methods (M
arXiv Detail & Related papers (2019-07-30T02:21:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.