Black-Box Min--Max Continuous Optimization Using CMA-ES with Worst-case
Ranking Approximation
- URL: http://arxiv.org/abs/2204.02646v1
- Date: Wed, 6 Apr 2022 08:03:39 GMT
- Title: Black-Box Min--Max Continuous Optimization Using CMA-ES with Worst-case
Ranking Approximation
- Authors: Atsuhiro Miyagi, Kazuto Fukuchi, Jun Sakuma, Youhei Akimoto
- Abstract summary: A popular approach updates $x$ and $y$ simultaneously or alternatingly.
Existing approaches fail if $f$ is not Lipschitz smooth and strongly convex-concave around the optimal solution.
We propose minimizing the worst-case objective function $F(x)=max_yf(x,y)$ directly using the covariance matrix adaptation evolution strategy.
- Score: 22.576922942465142
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In this study, we investigate the problem of min-max continuous optimization
in a black-box setting $\min_{x} \max_{y}f(x,y)$. A popular approach updates
$x$ and $y$ simultaneously or alternatingly. However, two major limitations
have been reported in existing approaches. (I) As the influence of the
interaction term between $x$ and $y$ (e.g., $x^\mathrm{T} B y$) on the
Lipschitz smooth and strongly convex-concave function $f$ increases, the
approaches converge to an optimal solution at a slower rate. (II) The
approaches fail to converge if $f$ is not Lipschitz smooth and strongly
convex-concave around the optimal solution. To address these difficulties, we
propose minimizing the worst-case objective function $F(x)=\max_{y}f(x,y)$
directly using the covariance matrix adaptation evolution strategy, in which
the rankings of solution candidates are approximated by our proposed worst-case
ranking approximation (WRA) mechanism. Compared with existing approaches,
numerical experiments show two important findings regarding our proposed
method. (1) The proposed approach is efficient in terms of $f$-calls on a
Lipschitz smooth and strongly convex-concave function with a large interaction
term. (2) The proposed approach can converge on functions that are not
Lipschitz smooth and strongly convex-concave around the optimal solution,
whereas existing approaches fail.
Related papers
- Methods for Convex $(L_0,L_1)$-Smooth Optimization: Clipping, Acceleration, and Adaptivity [50.25258834153574]
We focus on the class of (strongly) convex $(L0)$-smooth functions and derive new convergence guarantees for several existing methods.
In particular, we derive improved convergence rates for Gradient Descent with smoothnessed Gradient Clipping and for Gradient Descent with Polyak Stepsizes.
arXiv Detail & Related papers (2024-09-23T13:11:37Z) - Single-Loop Stochastic Algorithms for Difference of Max-Structured Weakly Convex Functions [41.43895948769255]
We show a class of non-smooth non-asymptotic fairness problems in the form of $min_x[yin Yphi(x, y) - max_zin Zpsix(x, z)]$.
We propose an envelope approximate gradient SMAG, the first method for solving these problems, provide a state-of-the-art non-asymptotic convergence rate.
arXiv Detail & Related papers (2024-05-28T20:52:46Z) - Nearly Minimax Optimal Regret for Learning Linear Mixture Stochastic
Shortest Path [80.60592344361073]
We study the Shortest Path (SSP) problem with a linear mixture transition kernel.
An agent repeatedly interacts with a environment and seeks to reach certain goal state while minimizing the cumulative cost.
Existing works often assume a strictly positive lower bound of the iteration cost function or an upper bound of the expected length for the optimal policy.
arXiv Detail & Related papers (2024-02-14T07:52:00Z) - An Accelerated Gradient Method for Convex Smooth Simple Bilevel Optimization [16.709026203727007]
We present a novel bilevel optimization method that locally approximates the solution set of the lower-level problem.
We measure the performance of our method in terms of suboptimality and infeasibility errors.
arXiv Detail & Related papers (2024-02-12T22:34:53Z) - Revisiting Inexact Fixed-Point Iterations for Min-Max Problems: Stochasticity and Structured Nonconvexity [18.427215139020632]
We show that $1L 1$ can be used to improve some state-of-the-art problems even for a multilevel Carlo iteration.
We provide an analysis for inexact Halperness estimators for $1L 1$ when the only hold with respect to a solution is a new $1L 1$ theory.
arXiv Detail & Related papers (2024-02-07T18:22:41Z) - An Oblivious Stochastic Composite Optimization Algorithm for Eigenvalue
Optimization Problems [76.2042837251496]
We introduce two oblivious mirror descent algorithms based on a complementary composite setting.
Remarkably, both algorithms work without prior knowledge of the Lipschitz constant or smoothness of the objective function.
We show how to extend our framework to scale and demonstrate the efficiency and robustness of our methods on large scale semidefinite programs.
arXiv Detail & Related papers (2023-06-30T08:34:29Z) - Covariance Matrix Adaptation Evolutionary Strategy with Worst-Case
Ranking Approximation for Min--Max Optimization and its Application to
Berthing Control Tasks [19.263468901608785]
We consider a continuous min--max optimization problem $min_x in mathbbX max_y in mathbbYf(x,y)$ whose objective function is a black-box.
We propose a novel approach to minimize the worst-case objective function $F(x) = max_y f(x,y)$ directly.
arXiv Detail & Related papers (2023-03-28T15:50:56Z) - Accelerated Primal-Dual Gradient Method for Smooth and Convex-Concave
Saddle-Point Problems with Bilinear Coupling [84.47780064014262]
We study a linear convex-concave saddle-point problem $min_xmax_y f(x) ytopmathbfA x - g(y)
arXiv Detail & Related papers (2021-12-30T20:31:46Z) - Minimax Optimization with Smooth Algorithmic Adversaries [59.47122537182611]
We propose a new algorithm for the min-player against smooth algorithms deployed by an adversary.
Our algorithm is guaranteed to make monotonic progress having no limit cycles, and to find an appropriate number of gradient ascents.
arXiv Detail & Related papers (2021-06-02T22:03:36Z) - Saddle Point Optimization with Approximate Minimization Oracle [8.680676599607125]
A major approach to saddle point optimization $min_xmax_y f(x, y)$ is a gradient based approach as is popularized by generative adversarial networks (GANs)
In contrast, we analyze an alternative approach relying only on an oracle that solves a minimization problem approximately.
Our approach locates approximate solutions $x'$ and $y'$ to $min_x'f(x', y)$ at a given point $(x, y)$ and updates $(x, y)$ toward these approximate solutions $(x', y'
arXiv Detail & Related papers (2021-03-29T23:03:24Z) - A Momentum-Assisted Single-Timescale Stochastic Approximation Algorithm
for Bilevel Optimization [112.59170319105971]
We propose a new algorithm -- the Momentum- Single-timescale Approximation (MSTSA) -- for tackling problems.
MSTSA allows us to control the error in iterations due to inaccurate solution to the lower level subproblem.
arXiv Detail & Related papers (2021-02-15T07:10:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.