Lower Bounds on the Worst-Case Complexity of Efficient Global
Optimization
- URL: http://arxiv.org/abs/2209.09655v1
- Date: Tue, 20 Sep 2022 11:57:13 GMT
- Title: Lower Bounds on the Worst-Case Complexity of Efficient Global
Optimization
- Authors: Wenjie Xu and Yuning Jiang and Emilio T. Maddalena and Colin N. Jones
- Abstract summary: We derive a unified lower bound for the complexity of efficient global optimization in terms of the metric entropy of a ball in its corresponding reproducing kernel Hilbert space(RKHS)
We show that this lower bound nearly matches the upper bound attained by non-adaptive search algorithms for the commonly used squared exponential kernel and the Mat'ern kernel.
- Score: 11.523746174066702
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Efficient global optimization is a widely used method for optimizing
expensive black-box functions such as tuning hyperparameter, and designing new
material, etc. Despite its popularity, less attention has been paid to
analyzing the inherent hardness of the problem although, given its extensive
use, it is important to understand the fundamental limits of efficient global
optimization algorithms. In this paper, we study the worst-case complexity of
the efficient global optimization problem and, in contrast to existing
kernel-specific results, we derive a unified lower bound for the complexity of
efficient global optimization in terms of the metric entropy of a ball in its
corresponding reproducing kernel Hilbert space~(RKHS). Specifically, we show
that if there exists a deterministic algorithm that achieves suboptimality gap
smaller than $\epsilon$ for any function $f\in S$ in $T$ function evaluations,
it is necessary that $T$ is at least
$\Omega\left(\frac{\log\mathcal{N}(S(\mathcal{X}),
4\epsilon,\|\cdot\|_\infty)}{\log(\frac{R}{\epsilon})}\right)$, where
$\mathcal{N}(\cdot,\cdot,\cdot)$ is the covering number, $S$ is the ball
centered at $0$ with radius $R$ in the RKHS and $S(\mathcal{X})$ is the
restriction of $S$ over the feasible set $\mathcal{X}$. Moreover, we show that
this lower bound nearly matches the upper bound attained by non-adaptive search
algorithms for the commonly used squared exponential kernel and the Mat\'ern
kernel with a large smoothness parameter $\nu$, up to a replacement of $d/2$ by
$d$ and a logarithmic term $\log\frac{R}{\epsilon}$. That is to say, our lower
bound is nearly optimal for these kernels.
Related papers
- Gaussian Process Bandit Optimization with Few Batches [49.896920704012395]
We introduce a batch algorithm inspired by finite-arm bandit algorithms.
We show that it achieves the cumulative regret upper bound $Oast(sqrtTgamma_T)$ using $O(loglog T)$ batches within time horizon $T$.
In addition, we propose a modified version of our algorithm, and characterize how the regret is impacted by the number of batches.
arXiv Detail & Related papers (2021-10-15T00:54:04Z) - Bayesian Optimistic Optimisation with Exponentially Decaying Regret [58.02542541410322]
The current practical BO algorithms have regret bounds ranging from $mathcalO(fraclogNsqrtN)$ to $mathcal O(e-sqrtN)$, where $N$ is the number of evaluations.
This paper explores the possibility of improving the regret bound in the noiseless setting by intertwining concepts from BO and tree-based optimistic optimisation.
We propose the BOO algorithm, a first practical approach which can achieve an exponential regret bound with order $mathcal O(N-sqrt
arXiv Detail & Related papers (2021-05-10T13:07:44Z) - Private Stochastic Convex Optimization: Optimal Rates in $\ell_1$
Geometry [69.24618367447101]
Up to logarithmic factors the optimal excess population loss of any $(varepsilon,delta)$-differently private is $sqrtlog(d)/n + sqrtd/varepsilon n.$
We show that when the loss functions satisfy additional smoothness assumptions, the excess loss is upper bounded (up to logarithmic factors) by $sqrtlog(d)/n + (log(d)/varepsilon n)2/3.
arXiv Detail & Related papers (2021-03-02T06:53:44Z) - Optimal Regret Algorithm for Pseudo-1d Bandit Convex Optimization [51.23789922123412]
We study online learning with bandit feedback (i.e. learner has access to only zeroth-order oracle) where cost/reward functions admit a "pseudo-1d" structure.
We show a lower bound of $min(sqrtdT, T3/4)$ for the regret of any algorithm, where $T$ is the number of rounds.
We propose a new algorithm sbcalg that combines randomized online gradient descent with a kernelized exponential weights method to exploit the pseudo-1d structure effectively.
arXiv Detail & Related papers (2021-02-15T08:16:51Z) - Finding Global Minima via Kernel Approximations [90.42048080064849]
We consider the global minimization of smooth functions based solely on function evaluations.
In this paper, we consider an approach that jointly models the function to approximate and finds a global minimum.
arXiv Detail & Related papers (2020-12-22T12:59:30Z) - Streaming Complexity of SVMs [110.63976030971106]
We study the space complexity of solving the bias-regularized SVM problem in the streaming model.
We show that for both problems, for dimensions of $frac1lambdaepsilon$, one can obtain streaming algorithms with spacely smaller than $frac1lambdaepsilon$.
arXiv Detail & Related papers (2020-07-07T17:10:00Z) - Multi-Scale Zero-Order Optimization of Smooth Functions in an RKHS [19.252319300590653]
A black-box function $f:mathcalX mapsto mathbbR$ is optimized under the assumption that $f$ is H"older smooth and has bounded norm in the RKHS associated with a given kernel $K$.
In this paper, we propose a new algorithm (textttLP-GP-UCB) where the usual GP surrogate model is augmented with Local Polynomial (LP) estimators of the H" smooth parameter $f$.
arXiv Detail & Related papers (2020-05-11T01:55:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.