Decentralized gradient descent maximization method for composite
nonconvex strongly-concave minimax problems
- URL: http://arxiv.org/abs/2304.02441v1
- Date: Wed, 5 Apr 2023 13:54:43 GMT
- Title: Decentralized gradient descent maximization method for composite
nonconvex strongly-concave minimax problems
- Authors: Yangyang Xu
- Abstract summary: We make the first attempt on solving NCSC minimax problems that can focus on both stationary nonsmooth terms.
Our algorithm is designed based on a novel reformulation of the minimax problem.
- Score: 7.5573375809946395
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Minimax problems have recently attracted a lot of research interests. A few
efforts have been made to solve decentralized nonconvex strongly-concave (NCSC)
minimax-structured optimization; however, all of them focus on smooth problems
with at most a constraint on the maximization variable. In this paper, we make
the first attempt on solving composite NCSC minimax problems that can have
convex nonsmooth terms on both minimization and maximization variables. Our
algorithm is designed based on a novel reformulation of the decentralized
minimax problem that introduces a multiplier to absorb the dual consensus
constraint. The removal of dual consensus constraint enables the most
aggressive (i.e., local maximization instead of a gradient ascent step) dual
update that leads to the benefit of taking a larger primal stepsize and better
complexity results. In addition, the decoupling of the nonsmoothness and
consensus on the dual variable eases the analysis of a decentralized algorithm;
thus our reformulation creates a new way for interested researchers to design
new (and possibly more efficient) decentralized methods on solving NCSC minimax
problems. We show a global convergence result of the proposed algorithm and an
iteration complexity result to produce a (near) stationary point of the
reformulation. Moreover, a relation is established between the (near)
stationarities of the reformulation and the original formulation. With this
relation, we show that when the dual regularizer is smooth, our algorithm can
have lower complexity results (with reduced dependence on a condition number)
than existing ones to produce a near-stationary point of the original
formulation. Numerical experiments are conducted on a distributionally robust
logistic regression to demonstrate the performance of the proposed algorithm.
Related papers
- Can Decentralized Stochastic Minimax Optimization Algorithms Converge
Linearly for Finite-Sum Nonconvex-Nonconcave Problems? [56.62372517641597]
Decentralized minimax optimization has been actively studied in the past few years due to its application in a wide range machine learning.
This paper develops two novel decentralized minimax optimization algorithms for the non-strongly-nonconcave problem.
arXiv Detail & Related papers (2023-04-24T02:19:39Z) - Decentralized Riemannian Algorithm for Nonconvex Minimax Problems [82.50374560598493]
The minimax algorithms for neural networks have been developed to solve many problems.
In this paper, we propose two types of minimax algorithms.
For the setting, we propose DRSGDA and prove that our method achieves a gradient.
arXiv Detail & Related papers (2023-02-08T01:42:45Z) - A Variance-Reduced Stochastic Gradient Tracking Algorithm for
Decentralized Optimization with Orthogonality Constraints [7.028225540638832]
We propose a novel algorithm for decentralized optimization with orthogonality constraints.
VRSGT is the first algorithm for decentralized optimization with orthogonality constraints that reduces both sampling and communication complexities simultaneously.
In the numerical experiments, VRGTS has a promising performance in a real-world autonomous sample.
arXiv Detail & Related papers (2022-08-29T14:46:44Z) - Minimax Optimization: The Case of Convex-Submodular [50.03984152441271]
Minimax problems extend beyond the continuous domain to mixed continuous-discrete domains or even fully discrete domains.
We introduce the class of convex-submodular minimax problems, where the objective is convex with respect to the continuous variable and submodular with respect to the discrete variable.
Our proposed algorithms are iterative and combine tools from both discrete and continuous optimization.
arXiv Detail & Related papers (2021-11-01T21:06:35Z) - Faster Algorithm and Sharper Analysis for Constrained Markov Decision
Process [56.55075925645864]
The problem of constrained decision process (CMDP) is investigated, where an agent aims to maximize the expected accumulated discounted reward subject to multiple constraints.
A new utilities-dual convex approach is proposed with novel integration of three ingredients: regularized policy, dual regularizer, and Nesterov's gradient descent dual.
This is the first demonstration that nonconcave CMDP problems can attain the lower bound of $mathcal O (1/epsilon)$ for all complexity optimization subject to convex constraints.
arXiv Detail & Related papers (2021-10-20T02:57:21Z) - Lower Bounds and Optimal Algorithms for Smooth and Strongly Convex
Decentralized Optimization Over Time-Varying Networks [79.16773494166644]
We consider the task of minimizing the sum of smooth and strongly convex functions stored in a decentralized manner across the nodes of a communication network.
We design two optimal algorithms that attain these lower bounds.
We corroborate the theoretical efficiency of these algorithms by performing an experimental comparison with existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-08T15:54:44Z) - An efficient nonconvex reformulation of stagewise convex optimization
problems [21.61123565780424]
We develop a non reformulation designed to exploit staged structure problems.
For neural network verification approach obtains spurious local gaps in only a few steps.
It can quickly solve problems faster than both offthe-shelf and specialized solvers.
arXiv Detail & Related papers (2020-10-27T14:30:32Z) - A Feasible Level Proximal Point Method for Nonconvex Sparse Constrained
Optimization [25.73397307080647]
We present a new model of a general convex or non objective machine machine objectives.
We propose an algorithm that solves a constraint with gradually relaxed point levels of each subproblem.
We demonstrate the effectiveness of our new numerical scale problems.
arXiv Detail & Related papers (2020-10-23T05:24:05Z) - Cogradient Descent for Bilinear Optimization [124.45816011848096]
We introduce a Cogradient Descent algorithm (CoGD) to address the bilinear problem.
We solve one variable by considering its coupling relationship with the other, leading to a synchronous gradient descent.
Our algorithm is applied to solve problems with one variable under the sparsity constraint.
arXiv Detail & Related papers (2020-06-16T13:41:54Z) - Stochastic gradient algorithms from ODE splitting perspective [0.0]
We present a different view on optimization, which goes back to the splitting schemes for approximate solutions of ODE.
In this work, we provide a connection between descent approach and gradient first-order splitting scheme for ODE.
We consider the special case of splitting, which is inspired by machine learning applications and derive a new upper bound on the global splitting error for it.
arXiv Detail & Related papers (2020-04-19T22:45:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.