Adaptive Bandit Convex Optimization with Heterogeneous Curvature
- URL: http://arxiv.org/abs/2202.06150v1
- Date: Sat, 12 Feb 2022 21:55:42 GMT
- Title: Adaptive Bandit Convex Optimization with Heterogeneous Curvature
- Authors: Haipeng Luo, Mengxiao Zhang, Peng Zhao
- Abstract summary: We study a heterogeneous setting where each function has its own curvature that is only revealed after the learner makes a decision.
We develop an efficient algorithm that is able to adapt to the curvature on the fly.
- Score: 40.368893108265056
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of adversarial bandit convex optimization, that is,
online learning over a sequence of arbitrary convex loss functions with only
one function evaluation for each of them. While all previous works assume known
and homogeneous curvature on these loss functions, we study a heterogeneous
setting where each function has its own curvature that is only revealed after
the learner makes a decision. We develop an efficient algorithm that is able to
adapt to the curvature on the fly. Specifically, our algorithm not only
recovers or \emph{even improves} existing results for several homogeneous
settings, but also leads to surprising results for some heterogeneous settings
-- for example, while Hazan and Levy (2014) showed that
$\widetilde{O}(d^{3/2}\sqrt{T})$ regret is achievable for a sequence of $T$
smooth and strongly convex $d$-dimensional functions, our algorithm reveals
that the same is achievable even if $T^{3/4}$ of them are not strongly convex,
and sometimes even if a constant fraction of them are not strongly convex. Our
approach is inspired by the framework of Bartlett et al. (2007) who studied a
similar heterogeneous setting but with stronger gradient feedback. Extending
their framework to the bandit feedback setting requires novel ideas such as
lifting the feasible domain and using a logarithmically homogeneous
self-concordant barrier regularizer.
Related papers
- Stochastic Zeroth-Order Optimization under Strongly Convexity and Lipschitz Hessian: Minimax Sample Complexity [59.75300530380427]
We consider the problem of optimizing second-order smooth and strongly convex functions where the algorithm is only accessible to noisy evaluations of the objective function it queries.
We provide the first tight characterization for the rate of the minimax simple regret by developing matching upper and lower bounds.
arXiv Detail & Related papers (2024-06-28T02:56:22Z) - Universal Online Learning with Gradient Variations: A Multi-layer Online Ensemble Approach [57.92727189589498]
We propose an online convex optimization approach with two different levels of adaptivity.
We obtain $mathcalO(log V_T)$, $mathcalO(d log V_T)$ and $hatmathcalO(sqrtV_T)$ regret bounds for strongly convex, exp-concave and convex loss functions.
arXiv Detail & Related papers (2023-07-17T09:55:35Z) - Gradient-free optimization of highly smooth functions: improved analysis
and a new algorithm [87.22224691317766]
This work studies problems with zero-order noisy oracle information under the assumption that the objective function is highly smooth.
We consider two kinds of zero-order projected gradient descent algorithms.
arXiv Detail & Related papers (2023-06-03T17:05:13Z) - Stochastic Nonsmooth Convex Optimization with Heavy-Tailed Noises:
High-Probability Bound, In-Expectation Rate and Initial Distance Adaptation [22.758674468435302]
In a heavy-tailed noise regime, the difference between the gradient and the true rate is assumed to have a finite $p-th moment.
This paper provides a comprehensive analysis of nonsmooth convex optimization with heavy-tailed noises.
arXiv Detail & Related papers (2023-03-22T03:05:28Z) - Randomized Coordinate Subgradient Method for Nonsmooth Composite
Optimization [11.017632675093628]
Coordinate-type subgradient methods for addressing nonsmooth problems are relatively underexplored due to the set of properties of the Lipschitz-type assumption.
arXiv Detail & Related papers (2022-06-30T02:17:11Z) - Optimal Extragradient-Based Bilinearly-Coupled Saddle-Point Optimization [116.89941263390769]
We consider the smooth convex-concave bilinearly-coupled saddle-point problem, $min_mathbfxmax_mathbfyF(mathbfx) + H(mathbfx,mathbfy)$, where one has access to first-order oracles for $F$, $G$ as well as the bilinear coupling function $H$.
We present a emphaccelerated gradient-extragradient (AG-EG) descent-ascent algorithm that combines extragrad
arXiv Detail & Related papers (2022-06-17T06:10:20Z) - Mirror Descent Strikes Again: Optimal Stochastic Convex Optimization
under Infinite Noise Variance [17.199063087458907]
We quantify the convergence rate of the Mirror Descent algorithm with a class of uniformly convex mirror maps.
This algorithm does not require any explicit gradient clipping or normalization.
We complement our results with information-theoretic lower bounds showing that no other algorithm using only first-order oracles can achieve improved rates.
arXiv Detail & Related papers (2022-02-23T17:08:40Z) - Optimal Rates for Random Order Online Optimization [60.011653053877126]
We study the citetgarber 2020online, where the loss functions may be chosen by an adversary, but are then presented online in a uniformly random order.
We show that citetgarber 2020online algorithms achieve the optimal bounds and significantly improve their stability.
arXiv Detail & Related papers (2021-06-29T09:48:46Z) - Provably Convergent Working Set Algorithm for Non-Convex Regularized
Regression [0.0]
This paper proposes a working set algorithm for non-regular regularizers with convergence guarantees.
Our results demonstrate high gain compared to the full problem solver for both block-coordinates or a gradient solver.
arXiv Detail & Related papers (2020-06-24T07:40:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.