Minimax Optimization: The Case of Convex-Submodular
- URL: http://arxiv.org/abs/2111.01262v1
- Date: Mon, 1 Nov 2021 21:06:35 GMT
- Title: Minimax Optimization: The Case of Convex-Submodular
- Authors: Arman Adibi, Aryan Mokhtari, Hamed Hassani
- Abstract summary: Minimax problems extend beyond the continuous domain to mixed continuous-discrete domains or even fully discrete domains.
We introduce the class of convex-submodular minimax problems, where the objective is convex with respect to the continuous variable and submodular with respect to the discrete variable.
Our proposed algorithms are iterative and combine tools from both discrete and continuous optimization.
- Score: 50.03984152441271
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Minimax optimization has been central in addressing various applications in
machine learning, game theory, and control theory. Prior literature has thus
far mainly focused on studying such problems in the continuous domain, e.g.,
convex-concave minimax optimization is now understood to a significant extent.
Nevertheless, minimax problems extend far beyond the continuous domain to mixed
continuous-discrete domains or even fully discrete domains. In this paper, we
study mixed continuous-discrete minimax problems where the minimization is over
a continuous variable belonging to Euclidean space and the maximization is over
subsets of a given ground set. We introduce the class of convex-submodular
minimax problems, where the objective is convex with respect to the continuous
variable and submodular with respect to the discrete variable. Even though such
problems appear frequently in machine learning applications, little is known
about how to address them from algorithmic and theoretical perspectives. For
such problems, we first show that obtaining saddle points are hard up to any
approximation, and thus introduce new notions of (near-) optimality. We then
provide several algorithmic procedures for solving convex and
monotone-submodular minimax problems and characterize their convergence rates,
computational complexity, and quality of the final solution according to our
notions of optimally. Our proposed algorithms are iterative and combine tools
from both discrete and continuous optimization. Finally, we provide numerical
experiments to showcase the effectiveness of our purposed methods.
Related papers
- Can Decentralized Stochastic Minimax Optimization Algorithms Converge
Linearly for Finite-Sum Nonconvex-Nonconcave Problems? [56.62372517641597]
Decentralized minimax optimization has been actively studied in the past few years due to its application in a wide range machine learning.
This paper develops two novel decentralized minimax optimization algorithms for the non-strongly-nonconcave problem.
arXiv Detail & Related papers (2023-04-24T02:19:39Z) - Decentralized gradient descent maximization method for composite
nonconvex strongly-concave minimax problems [7.5573375809946395]
We make the first attempt on solving NCSC minimax problems that can focus on both stationary nonsmooth terms.
Our algorithm is designed based on a novel reformulation of the minimax problem.
arXiv Detail & Related papers (2023-04-05T13:54:43Z) - Escaping limit cycles: Global convergence for constrained
nonconvex-nonconcave minimax problems [46.71914490521821]
This paper introduces a new extragradient-type algorithm for a class of non-nonconcave minimax problems.
The proposed algorithm is applicable to constrained and regularized problems.
arXiv Detail & Related papers (2023-02-20T08:37:08Z) - Decentralized Riemannian Algorithm for Nonconvex Minimax Problems [82.50374560598493]
The minimax algorithms for neural networks have been developed to solve many problems.
In this paper, we propose two types of minimax algorithms.
For the setting, we propose DRSGDA and prove that our method achieves a gradient.
arXiv Detail & Related papers (2023-02-08T01:42:45Z) - A Decentralized Adaptive Momentum Method for Solving a Class of Min-Max
Optimization Problems [9.653157073271021]
We develop a decentralized adaptive momentum (ADAM)-type algorithm for solving min-max optimization problem.
We obtain non-asymptotic rates of convergence of the proposed algorithm for finding a (stochastic) first-order Nash equilibrium point.
arXiv Detail & Related papers (2021-06-10T22:32:01Z) - Efficient Methods for Structured Nonconvex-Nonconcave Min-Max
Optimization [98.0595480384208]
We propose a generalization extraient spaces which converges to a stationary point.
The algorithm applies not only to general $p$-normed spaces, but also to general $p$-dimensional vector spaces.
arXiv Detail & Related papers (2020-10-31T21:35:42Z) - Gradient Free Minimax Optimization: Variance Reduction and Faster
Convergence [120.9336529957224]
In this paper, we denote the non-strongly setting on the magnitude of a gradient-free minimax optimization problem.
We show that a novel zeroth-order variance reduced descent algorithm achieves the best known query complexity.
arXiv Detail & Related papers (2020-06-16T17:55:46Z) - Non-convex Min-Max Optimization: Applications, Challenges, and Recent
Theoretical Advances [58.54078318403909]
The min-max problem, also known as the saddle point problem, is a class adversarial problem which is also studied in the context ofsum games.
arXiv Detail & Related papers (2020-06-15T05:33:42Z) - Halpern Iteration for Near-Optimal and Parameter-Free Monotone Inclusion
and Strong Solutions to Variational Inequalities [14.848525762485872]
We leverage the connections between nonexpansive maps, monotone Lipschitz operators, and proximal mappings to obtain near-optimal solutions to monotone inclusion problems.
These results translate into near-optimal guarantees for approximating strong solutions to variational inequality problems, approximating convex-concave min-max optimization problems, and minimizing the norm of the gradient in min-max optimization problems.
arXiv Detail & Related papers (2020-02-20T17:12:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.