Near-linear Time Gaussian Process Optimization with Adaptive Batching
and Resparsification
- URL: http://arxiv.org/abs/2002.09954v2
- Date: Wed, 26 Feb 2020 09:45:29 GMT
- Title: Near-linear Time Gaussian Process Optimization with Adaptive Batching
and Resparsification
- Authors: Daniele Calandriello, Luigi Carratino, Alessandro Lazaric, Michal
Valko, Lorenzo Rosasco
- Abstract summary: We introduce BBKB, the first no-regret GP optimization algorithm that provably runs in near-linear time and selects candidates in batches.
We show that the same bound can be used to adaptively delay costly updates to the sparse GP approximation, achieving a near-constant per-step amortized cost.
- Score: 119.41129787351092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gaussian processes (GP) are one of the most successful frameworks to model
uncertainty. However, GP optimization (e.g., GP-UCB) suffers from major
scalability issues. Experimental time grows linearly with the number of
evaluations, unless candidates are selected in batches (e.g., using GP-BUCB)
and evaluated in parallel. Furthermore, computational cost is often prohibitive
since algorithms such as GP-BUCB require a time at least quadratic in the
number of dimensions and iterations to select each batch. In this paper, we
introduce BBKB (Batch Budgeted Kernel Bandits), the first no-regret GP
optimization algorithm that provably runs in near-linear time and selects
candidates in batches. This is obtained with a new guarantee for the tracking
of the posterior variances that allows BBKB to choose increasingly larger
batches, improving over GP-BUCB. Moreover, we show that the same bound can be
used to adaptively delay costly updates to the sparse GP approximation used by
BBKB, achieving a near-constant per-step amortized cost. These findings are
then confirmed in several experiments, where BBKB is much faster than
state-of-the-art methods.
Related papers
- Sample-Then-Optimize Batch Neural Thompson Sampling [50.800944138278474]
We introduce two algorithms for black-box optimization based on the Thompson sampling (TS) policy.
To choose an input query, we only need to train an NN and then choose the query by maximizing the trained NN.
Our algorithms sidestep the need to invert the large parameter matrix yet still preserve the validity of the TS policy.
arXiv Detail & Related papers (2022-10-13T09:01:58Z) - Fast Gaussian Process Posterior Mean Prediction via Local Cross
Validation and Precomputation [0.0]
We present a fast posterior mean prediction algorithm called FastMuyGPs.
It is based upon the MuyGPs hyper parameter estimation algorithm and utilizes a combination of leave-one-out cross-validation, nearest neighbors sparsification, and precomputation.
It attains superior accuracy and competitive or superior runtime to both deep neural networks and state-of-the-art GP algorithms.
arXiv Detail & Related papers (2022-05-22T17:38:36Z) - Scaling Gaussian Process Optimization by Evaluating a Few Unique
Candidates Multiple Times [119.41129787351092]
We show that sequential black-box optimization based on GPs can be made efficient by sticking to a candidate solution for multiple evaluation steps.
We modify two well-established GP-Opt algorithms, GP-UCB and GP-EI to adapt rules from batched GP-Opt.
arXiv Detail & Related papers (2022-01-30T20:42:14Z) - Misspecified Gaussian Process Bandit Optimization [59.30399661155574]
Kernelized bandit algorithms have shown strong empirical and theoretical performance for this problem.
We introduce a emphmisspecified kernelized bandit setting where the unknown function can be $epsilon$--uniformly approximated by a function with a bounded norm in some Reproducing Kernel Hilbert Space (RKHS)
We show that our algorithm achieves optimal dependence on $epsilon$ with no prior knowledge of misspecification.
arXiv Detail & Related papers (2021-11-09T09:00:02Z) - Fast and Scalable Spike and Slab Variable Selection in High-Dimensional
Gaussian Processes [12.667478571732449]
We develop a fast and scalable variational inference algorithm for the spike and slab GP that is tractable with arbitrary differentiable kernels.
In experiments our method consistently outperforms vanilla and sparse variational GPs whilst retaining similar runtimes.
arXiv Detail & Related papers (2021-11-08T15:13:24Z) - MuyGPs: Scalable Gaussian Process Hyperparameter Estimation Using Local
Cross-Validation [1.2233362977312945]
We present MuyGPs, a novel efficient GP hyper parameter estimation method.
MuyGPs builds upon prior methods that take advantage of the nearest neighbors structure of the data.
We show that our method outperforms all known competitors both in terms of time-to-solution and the root mean squared error of the predictions.
arXiv Detail & Related papers (2021-04-29T18:10:21Z) - No-Regret Algorithms for Time-Varying Bayesian Optimization [0.0]
We adopt the general variation budget model to capture the time-varying environment.
We introduce two GP-UCB type algorithms: R-GP-UCB and SW-GP-UCB, respectively.
Our results not only recover previous linear bandit results when a linear kernel is used, but complement the previous regret analysis of time-varying Gaussian process bandit.
arXiv Detail & Related papers (2021-02-11T22:35:32Z) - Likelihood-Free Inference with Deep Gaussian Processes [70.74203794847344]
Surrogate models have been successfully used in likelihood-free inference to decrease the number of simulator evaluations.
We propose a Deep Gaussian Process (DGP) surrogate model that can handle more irregularly behaved target distributions.
Our experiments show how DGPs can outperform GPs on objective functions with multimodal distributions and maintain a comparable performance in unimodal cases.
arXiv Detail & Related papers (2020-06-18T14:24:05Z) - Uncertainty quantification using martingales for misspecified Gaussian
processes [52.22233158357913]
We address uncertainty quantification for Gaussian processes (GPs) under misspecified priors.
We construct a confidence sequence (CS) for the unknown function using martingale techniques.
Our CS is statistically valid and empirically outperforms standard GP methods.
arXiv Detail & Related papers (2020-06-12T17:58:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.