Nearly Optimal Algorithms for Level Set Estimation
- URL: http://arxiv.org/abs/2111.01768v1
- Date: Tue, 2 Nov 2021 17:45:02 GMT
- Title: Nearly Optimal Algorithms for Level Set Estimation
- Authors: Blake Mason, Romain Camilleri, Subhojyoti Mukherjee, Kevin Jamieson,
Robert Nowak, Lalit Jain
- Abstract summary: We provide a new approach to the level set estimation problem by relating it to recent adaptive experimental design methods for linear bandits.
We show that our bounds are nearly optimal, namely, our upper bounds match existing lower bounds for threshold linear bandits.
- Score: 21.83736847203543
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The level set estimation problem seeks to find all points in a domain ${\cal
X}$ where the value of an unknown function $f:{\cal X}\rightarrow \mathbb{R}$
exceeds a threshold $\alpha$. The estimation is based on noisy function
evaluations that may be acquired at sequentially and adaptively chosen
locations in ${\cal X}$. The threshold value $\alpha$ can either be
\emph{explicit} and provided a priori, or \emph{implicit} and defined relative
to the optimal function value, i.e. $\alpha = (1-\epsilon)f(x_\ast)$ for a
given $\epsilon > 0$ where $f(x_\ast)$ is the maximal function value and is
unknown. In this work we provide a new approach to the level set estimation
problem by relating it to recent adaptive experimental design methods for
linear bandits in the Reproducing Kernel Hilbert Space (RKHS) setting. We
assume that $f$ can be approximated by a function in the RKHS up to an unknown
misspecification and provide novel algorithms for both the implicit and
explicit cases in this setting with strong theoretical guarantees. Moreover, in
the linear (kernel) setting, we show that our bounds are nearly optimal,
namely, our upper bounds match existing lower bounds for threshold linear
bandits. To our knowledge this work provides the first instance-dependent,
non-asymptotic upper bounds on sample complexity of level-set estimation that
match information theoretic lower bounds.
Related papers
- Nearly Minimax Optimal Regret for Learning Linear Mixture Stochastic
Shortest Path [80.60592344361073]
We study the Shortest Path (SSP) problem with a linear mixture transition kernel.
An agent repeatedly interacts with a environment and seeks to reach certain goal state while minimizing the cumulative cost.
Existing works often assume a strictly positive lower bound of the iteration cost function or an upper bound of the expected length for the optimal policy.
arXiv Detail & Related papers (2024-02-14T07:52:00Z) - High Probability Convergence of Stochastic Gradient Methods [15.829413808059124]
We show convergence with bounds depending on the initial distance to the optimal solution.
We demonstrate that our techniques can be used to obtain high bound for AdaGrad-Norm.
arXiv Detail & Related papers (2023-02-28T18:42:11Z) - Estimating the minimizer and the minimum value of a regression function
under passive design [72.85024381807466]
We propose a new method for estimating the minimizer $boldsymbolx*$ and the minimum value $f*$ of a smooth and strongly convex regression function $f$.
We derive non-asymptotic upper bounds for the quadratic risk and optimization error of $boldsymbolz_n$, and for the risk of estimating $f*$.
arXiv Detail & Related papers (2022-11-29T18:38:40Z) - Best Policy Identification in Linear MDPs [70.57916977441262]
We investigate the problem of best identification in discounted linear Markov+Delta Decision in the fixed confidence setting under a generative model.
The lower bound as the solution of an intricate non- optimization program can be used as the starting point to devise such algorithms.
arXiv Detail & Related papers (2022-08-11T04:12:50Z) - A first-order primal-dual method with adaptivity to local smoothness [64.62056765216386]
We consider the problem of finding a saddle point for the convex-concave objective $min_x max_y f(x) + langle Ax, yrangle - g*(y)$, where $f$ is a convex function with locally Lipschitz gradient and $g$ is convex and possibly non-smooth.
We propose an adaptive version of the Condat-Vu algorithm, which alternates between primal gradient steps and dual steps.
arXiv Detail & Related papers (2021-10-28T14:19:30Z) - Second-Order Information in Non-Convex Stochastic Optimization: Power
and Limitations [54.42518331209581]
We find an algorithm which finds.
epsilon$-approximate stationary point (with $|nabla F(x)|le epsilon$) using.
$(epsilon,gamma)$surimate random random points.
Our lower bounds here are novel even in the noiseless case.
arXiv Detail & Related papers (2020-06-24T04:41:43Z) - Multi-Scale Zero-Order Optimization of Smooth Functions in an RKHS [19.252319300590653]
A black-box function $f:mathcalX mapsto mathbbR$ is optimized under the assumption that $f$ is H"older smooth and has bounded norm in the RKHS associated with a given kernel $K$.
In this paper, we propose a new algorithm (textttLP-GP-UCB) where the usual GP surrogate model is augmented with Local Polynomial (LP) estimators of the H" smooth parameter $f$.
arXiv Detail & Related papers (2020-05-11T01:55:39Z) - Complexity of Finding Stationary Points of Nonsmooth Nonconvex Functions [84.49087114959872]
We provide the first non-asymptotic analysis for finding stationary points of nonsmooth, nonsmooth functions.
In particular, we study Hadamard semi-differentiable functions, perhaps the largest class of nonsmooth functions.
arXiv Detail & Related papers (2020-02-10T23:23:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.