Conic Blackwell Algorithm: Parameter-Free Convex-Concave Saddle-Point
Solving
- URL: http://arxiv.org/abs/2105.13203v1
- Date: Thu, 27 May 2021 14:50:31 GMT
- Title: Conic Blackwell Algorithm: Parameter-Free Convex-Concave Saddle-Point
Solving
- Authors: Julien Grand-Cl\'ement, Christian Kroer
- Abstract summary: We develop a new simple regret minimizer, the Conic Blackwell Algorithm$+$ (CBA$+$)
We show how to implement CBA$+$ for the simplex, $ell_p$ norm balls, and ellipsoidal confidence regions in the simplex.
We present numerical experiments for solving matrix games and distributionally robust optimization problems.
- Score: 21.496093093434357
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We develop new parameter and scale-free algorithms for solving convex-concave
saddle-point problems. Our results are based on a new simple regret minimizer,
the Conic Blackwell Algorithm$^+$ (CBA$^+$), which attains $O(1/\sqrt{T})$
average regret. Intuitively, our approach generalizes to other decision sets of
interest ideas from the Counterfactual Regret minimization (CFR$^+$) algorithm,
which has very strong practical performance for solving sequential games on
simplexes. We show how to implement CBA$^+$ for the simplex, $\ell_{p}$ norm
balls, and ellipsoidal confidence regions in the simplex, and we present
numerical experiments for solving matrix games and distributionally robust
optimization problems. Our empirical results show that CBA$^+$ is a simple
algorithm that outperforms state-of-the-art methods on synthetic data and real
data instances, without the need for any choice of step sizes or other
algorithmic parameters.
Related papers
- Deterministic Nonsmooth Nonconvex Optimization [94.01526844386977]
We show that randomization is necessary to obtain a dimension-free dimension-free algorithm.
Our algorithm yields the first deterministic dimension-free algorithm for optimizing ReLU networks.
arXiv Detail & Related papers (2023-02-16T13:57:19Z) - Best Policy Identification in Linear MDPs [70.57916977441262]
We investigate the problem of best identification in discounted linear Markov+Delta Decision in the fixed confidence setting under a generative model.
The lower bound as the solution of an intricate non- optimization program can be used as the starting point to devise such algorithms.
arXiv Detail & Related papers (2022-08-11T04:12:50Z) - Asymptotic Instance-Optimal Algorithms for Interactive Decision Making [35.76184529520015]
In this paper, we design the first instance-optimal algorithm for general interactive decision making problems with finite number of decisions under mild conditions.
Our results, instantiated on concrete problems, recover the classical gap-dependent bounds for multi-armed bandits and prior works on linear bandits.
arXiv Detail & Related papers (2022-06-06T02:56:10Z) - Solving optimization problems with Blackwell approachability [31.29341288414507]
Conic Blackwell Algorithm$+$ (CBA$+$) regret minimizer.
CBA$+$ is based on Blackwell approachability and attains $O(sqrtT)$ regret.
Based on CBA$+$, we introduce SP-CBA$+$, a new parameter-free algorithm for solving convex-concave saddle-point problems.
arXiv Detail & Related papers (2022-02-24T18:19:21Z) - Derivative-free Alternating Projection Algorithms for General
Nonconvex-Concave Minimax Problems [9.173866646584031]
In this paper, we propose an algorithm for nonsmooth zeroth-order minimax problems.
We show that it can be used to attack nonconcave minimax problems.
arXiv Detail & Related papers (2021-08-01T15:23:49Z) - Randomized Exploration for Reinforcement Learning with General Value
Function Approximation [122.70803181751135]
We propose a model-free reinforcement learning algorithm inspired by the popular randomized least squares value iteration (RLSVI) algorithm.
Our algorithm drives exploration by simply perturbing the training data with judiciously chosen i.i.d. scalar noises.
We complement the theory with an empirical evaluation across known difficult exploration tasks.
arXiv Detail & Related papers (2021-06-15T02:23:07Z) - Minimax Optimization with Smooth Algorithmic Adversaries [59.47122537182611]
We propose a new algorithm for the min-player against smooth algorithms deployed by an adversary.
Our algorithm is guaranteed to make monotonic progress having no limit cycles, and to find an appropriate number of gradient ascents.
arXiv Detail & Related papers (2021-06-02T22:03:36Z) - Bayesian Optimistic Optimisation with Exponentially Decaying Regret [58.02542541410322]
The current practical BO algorithms have regret bounds ranging from $mathcalO(fraclogNsqrtN)$ to $mathcal O(e-sqrtN)$, where $N$ is the number of evaluations.
This paper explores the possibility of improving the regret bound in the noiseless setting by intertwining concepts from BO and tree-based optimistic optimisation.
We propose the BOO algorithm, a first practical approach which can achieve an exponential regret bound with order $mathcal O(N-sqrt
arXiv Detail & Related papers (2021-05-10T13:07:44Z) - Saddle Point Optimization with Approximate Minimization Oracle [8.680676599607125]
A major approach to saddle point optimization $min_xmax_y f(x, y)$ is a gradient based approach as is popularized by generative adversarial networks (GANs)
In contrast, we analyze an alternative approach relying only on an oracle that solves a minimization problem approximately.
Our approach locates approximate solutions $x'$ and $y'$ to $min_x'f(x', y)$ at a given point $(x, y)$ and updates $(x, y)$ toward these approximate solutions $(x', y'
arXiv Detail & Related papers (2021-03-29T23:03:24Z) - Efficient Projection-Free Algorithms for Saddle Point Problems [39.88460595129901]
We study projection-free algorithms for convex-strongly-concave saddle point problems with complicated constraints.
Our method combines Conditional Gradient Sliding with Mirror-Prox and shows that it only requires $tildeO (1/sqrtepsilon)$ evaluations and $tildeO (1/epsilon2)$ linear optimizations in the batch setting.
arXiv Detail & Related papers (2020-10-21T15:05:53Z) - Gradient Free Minimax Optimization: Variance Reduction and Faster
Convergence [120.9336529957224]
In this paper, we denote the non-strongly setting on the magnitude of a gradient-free minimax optimization problem.
We show that a novel zeroth-order variance reduced descent algorithm achieves the best known query complexity.
arXiv Detail & Related papers (2020-06-16T17:55:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.