Conditioning Sparse Variational Gaussian Processes for Online
Decision-making
- URL: http://arxiv.org/abs/2110.15172v1
- Date: Thu, 28 Oct 2021 14:49:34 GMT
- Title: Conditioning Sparse Variational Gaussian Processes for Online
Decision-making
- Authors: Wesley J. Maddox, Samuel Stanton, Andrew Gordon Wilson
- Abstract summary: Online variational conditioning (OVC) is a procedure for efficiently conditioning SVGPs in an online setting.
OVC provides compelling performance in a range of applications including active learning of malaria incidence.
- Score: 40.13509498161087
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With a principled representation of uncertainty and closed form posterior
updates, Gaussian processes (GPs) are a natural choice for online decision
making. However, Gaussian processes typically require at least
$\mathcal{O}(n^2)$ computations for $n$ training points, limiting their general
applicability. Stochastic variational Gaussian processes (SVGPs) can provide
scalable inference for a dataset of fixed size, but are difficult to
efficiently condition on new data. We propose online variational conditioning
(OVC), a procedure for efficiently conditioning SVGPs in an online setting that
does not require re-training through the evidence lower bound with the addition
of new data. OVC enables the pairing of SVGPs with advanced look-ahead
acquisition functions for black-box optimization, even with non-Gaussian
likelihoods. We show OVC provides compelling performance in a range of
applications including active learning of malaria incidence, and reinforcement
learning on MuJoCo simulated robotic control tasks.
Related papers
- Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference [55.150117654242706]
We show that model selection for computation-aware GPs trained on 1.8 million data points can be done within a few hours on a single GPU.
As a result of this work, Gaussian processes can be trained on large-scale datasets without significantly compromising their ability to quantify uncertainty.
arXiv Detail & Related papers (2024-11-01T21:11:48Z) - Approximation-Aware Bayesian Optimization [34.56666383247348]
High-dimensional Bayesian optimization (BO) tasks often require 10,000 function evaluations before obtaining meaningful results.
We modify sparse variational Gaussian processes (SVGPs) to better align with the goals of BO.
Using the framework of utility-calibrated variational inference, we unify GP approximation and data acquisition into a joint optimization problem.
arXiv Detail & Related papers (2024-06-06T17:55:02Z) - Closed-form Filtering for Non-linear Systems [83.91296397912218]
We propose a new class of filters based on Gaussian PSD Models, which offer several advantages in terms of density approximation and computational efficiency.
We show that filtering can be efficiently performed in closed form when transitions and observations are Gaussian PSD Models.
Our proposed estimator enjoys strong theoretical guarantees, with estimation error that depends on the quality of the approximation and is adaptive to the regularity of the transition probabilities.
arXiv Detail & Related papers (2024-02-15T08:51:49Z) - Revisiting Active Sets for Gaussian Process Decoders [0.0]
We develop a new estimate of the log-marginal likelihood based on recently discovered links to cross-validation.
We demonstrate that the resulting active sets (SAS) approximation significantly improves the robustness of GP decoder training.
arXiv Detail & Related papers (2022-09-10T10:49:31Z) - A piece-wise constant approximation for non-conjugate Gaussian Process
models [0.0]
This paper proposes to approximate the inverse-link function, which is necessary when working with non-Gaussian likelihoods.
It will be shown that this yields a closed form solution for the corresponding SVGP lower bound.
In addition, it is demonstrated how the piece-wise constant function itself can be optimized, resulting in an inverse-link function that can be learnt from the data at hand.
arXiv Detail & Related papers (2022-04-22T08:53:54Z) - An Experimental Design Perspective on Model-Based Reinforcement Learning [73.37942845983417]
In practical applications of RL, it is expensive to observe state transitions from the environment.
We propose an acquisition function that quantifies how much information a state-action pair would provide about the optimal solution to a Markov decision process.
arXiv Detail & Related papers (2021-12-09T23:13:57Z) - Kernel Stein Generative Modeling [68.03537693810972]
Gradient Langevin Dynamics (SGLD) demonstrates impressive results with energy-based models on high-dimensional and complex data distributions.
Stein Variational Gradient Descent (SVGD) is a deterministic sampling algorithm that iteratively transports a set of particles to approximate a given distribution.
We propose noise conditional kernel SVGD (NCK-SVGD), that works in tandem with the recently introduced Noise Conditional Score Network estimator.
arXiv Detail & Related papers (2020-07-06T21:26:04Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z) - $\pi$VAE: a stochastic process prior for Bayesian deep learning with
MCMC [2.4792948967354236]
We propose a novel variational autoencoder called the prior encodingal autoencoder ($pi$VAE)
We show that our framework can accurately learn expressive function classes such as Gaussian processes, but also properties of functions to enable statistical inference.
Perhaps most usefully, we demonstrate that the low dimensional distributed latent space representation learnt provides an elegant and scalable means of performing inference for processes within programming languages such as Stan.
arXiv Detail & Related papers (2020-02-17T10:23:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.