Multi-Scale Zero-Order Optimization of Smooth Functions in an RKHS
- URL: http://arxiv.org/abs/2005.04832v1
- Date: Mon, 11 May 2020 01:55:39 GMT
- Title: Multi-Scale Zero-Order Optimization of Smooth Functions in an RKHS
- Authors: Shubhanshu Shekhar, Tara Javidi
- Abstract summary: A black-box function $f:mathcalX mapsto mathbbR$ is optimized under the assumption that $f$ is H"older smooth and has bounded norm in the RKHS associated with a given kernel $K$.
In this paper, we propose a new algorithm (textttLP-GP-UCB) where the usual GP surrogate model is augmented with Local Polynomial (LP) estimators of the H" smooth parameter $f$.
- Score: 19.252319300590653
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We aim to optimize a black-box function $f:\mathcal{X} \mapsto \mathbb{R}$
under the assumption that $f$ is H\"older smooth and has bounded norm in the
RKHS associated with a given kernel $K$. This problem is known to have an
agnostic Gaussian Process (GP) bandit interpretation in which an appropriately
constructed GP surrogate model with kernel $K$ is used to obtain an upper
confidence bound (UCB) algorithm. In this paper, we propose a new algorithm
(\texttt{LP-GP-UCB}) where the usual GP surrogate model is augmented with Local
Polynomial (LP) estimators of the H\"older smooth function $f$ to construct a
multi-scale UCB guiding the search for the optimizer. We analyze this algorithm
and derive high probability bounds on its simple and cumulative regret. We then
prove that the elements of many common RKHS are H\"older smooth and obtain the
corresponding H\"older smoothness parameters, and hence, specialize our regret
bounds for several commonly used kernels. When specialized to the Squared
Exponential (SE) kernel, \texttt{LP-GP-UCB} matches the optimal performance,
while for the case of Mat\'ern kernels $(K_{\nu})_{\nu>0}$, it results in
uniformly tighter regret bounds for all values of the smoothness parameter
$\nu>0$. Most notably, for certain ranges of $\nu$, the algorithm achieves
near-optimal bounds on simple and cumulative regrets, matching the
algorithm-independent lower bounds up to polylog factors, and thus closing the
large gap between the existing upper and lower bounds for these values of
$\nu$. Additionally, our analysis provides the first explicit regret bounds, in
terms of the budget $n$, for the Rational-Quadratic (RQ) and Gamma-Exponential
(GE). Finally, experiments with synthetic functions as well as a CNN
hyperparameter tuning task demonstrate the practical benefits of our
multi-scale partitioning approach over some existing algorithms numerically.
Related papers
- Nearly Minimax Optimal Regret for Learning Linear Mixture Stochastic
Shortest Path [80.60592344361073]
We study the Shortest Path (SSP) problem with a linear mixture transition kernel.
An agent repeatedly interacts with a environment and seeks to reach certain goal state while minimizing the cumulative cost.
Existing works often assume a strictly positive lower bound of the iteration cost function or an upper bound of the expected length for the optimal policy.
arXiv Detail & Related papers (2024-02-14T07:52:00Z) - On Convergence of Incremental Gradient for Non-Convex Smooth Functions [63.51187646914962]
In machine learning and network optimization, algorithms like shuffle SGD are popular due to minimizing the number of misses and good cache.
This paper delves into the convergence properties SGD algorithms with arbitrary data ordering.
arXiv Detail & Related papers (2023-05-30T17:47:27Z) - Lower Bounds on the Worst-Case Complexity of Efficient Global
Optimization [11.523746174066702]
We derive a unified lower bound for the complexity of efficient global optimization in terms of the metric entropy of a ball in its corresponding reproducing kernel Hilbert space(RKHS)
We show that this lower bound nearly matches the upper bound attained by non-adaptive search algorithms for the commonly used squared exponential kernel and the Mat'ern kernel.
arXiv Detail & Related papers (2022-09-20T11:57:13Z) - Best Policy Identification in Linear MDPs [70.57916977441262]
We investigate the problem of best identification in discounted linear Markov+Delta Decision in the fixed confidence setting under a generative model.
The lower bound as the solution of an intricate non- optimization program can be used as the starting point to devise such algorithms.
arXiv Detail & Related papers (2022-08-11T04:12:50Z) - Misspecified Gaussian Process Bandit Optimization [59.30399661155574]
Kernelized bandit algorithms have shown strong empirical and theoretical performance for this problem.
We introduce a emphmisspecified kernelized bandit setting where the unknown function can be $epsilon$--uniformly approximated by a function with a bounded norm in some Reproducing Kernel Hilbert Space (RKHS)
We show that our algorithm achieves optimal dependence on $epsilon$ with no prior knowledge of misspecification.
arXiv Detail & Related papers (2021-11-09T09:00:02Z) - Gaussian Process Bandit Optimization with Few Batches [49.896920704012395]
We introduce a batch algorithm inspired by finite-arm bandit algorithms.
We show that it achieves the cumulative regret upper bound $Oast(sqrtTgamma_T)$ using $O(loglog T)$ batches within time horizon $T$.
In addition, we propose a modified version of our algorithm, and characterize how the regret is impacted by the number of batches.
arXiv Detail & Related papers (2021-10-15T00:54:04Z) - Optimal Order Simple Regret for Gaussian Process Bandits [6.84224661918159]
We show that a pure exploration algorithm is significantly tighter than the existing bounds.
We show that this regret is optimal up to logarithmically for cases where a lower bound on a kernel is known.
arXiv Detail & Related papers (2021-08-20T16:49:32Z) - Bayesian Optimistic Optimisation with Exponentially Decaying Regret [58.02542541410322]
The current practical BO algorithms have regret bounds ranging from $mathcalO(fraclogNsqrtN)$ to $mathcal O(e-sqrtN)$, where $N$ is the number of evaluations.
This paper explores the possibility of improving the regret bound in the noiseless setting by intertwining concepts from BO and tree-based optimistic optimisation.
We propose the BOO algorithm, a first practical approach which can achieve an exponential regret bound with order $mathcal O(N-sqrt
arXiv Detail & Related papers (2021-05-10T13:07:44Z) - On Information Gain and Regret Bounds in Gaussian Process Bandits [8.499604625449907]
Consider the optimization of an expensive to evaluate and possibly non- sequential objective function $f$ from noisy feedback observations.
For the Matern family of kernels, lower bounds on $gamma_T$, and regret under the frequentist setting, are known.
arXiv Detail & Related papers (2020-09-15T10:15:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.