Revisiting Active Sets for Gaussian Process Decoders
- URL: http://arxiv.org/abs/2209.04636v1
- Date: Sat, 10 Sep 2022 10:49:31 GMT
- Title: Revisiting Active Sets for Gaussian Process Decoders
- Authors: Pablo Moreno-Mu\~noz, Cilie W Feldager, S{\o}ren Hauberg
- Abstract summary: We develop a new estimate of the log-marginal likelihood based on recently discovered links to cross-validation.
We demonstrate that the resulting active sets (SAS) approximation significantly improves the robustness of GP decoder training.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Decoders built on Gaussian processes (GPs) are enticing due to the
marginalisation over the non-linear function space. Such models (also known as
GP-LVMs) are often expensive and notoriously difficult to train in practice,
but can be scaled using variational inference and inducing points. In this
paper, we revisit active set approximations. We develop a new stochastic
estimate of the log-marginal likelihood based on recently discovered links to
cross-validation, and propose a computationally efficient approximation
thereof. We demonstrate that the resulting stochastic active sets (SAS)
approximation significantly improves the robustness of GP decoder training
while reducing computational cost. The SAS-GP obtains more structure in the
latent space, scales to many datapoints and learns better representations than
variational autoencoders, which is rarely the case for GP decoders.
Related papers
- Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of Language Model [89.8764435351222]
We propose a new family of unbiased estimators called WTA-CRS, for matrix production with reduced variance.
Our work provides both theoretical and experimental evidence that, in the context of tuning transformers, our proposed estimators exhibit lower variance compared to existing ones.
arXiv Detail & Related papers (2023-05-24T15:52:08Z) - Actually Sparse Variational Gaussian Processes [20.71289963037696]
We propose a new class of inter-domain variational GP constructed by projecting a GP onto a set of compactly supported B-spline basis functions.
This allows us to very efficiently model fast-varying spatial phenomena with tens of thousands of inducing variables.
arXiv Detail & Related papers (2023-04-11T09:38:58Z) - Markovian Gaussian Process Variational Autoencoders [19.686719654642392]
We leverage the equivalent discrete state space representation of Markovian GPs to enable linear time GPVAE training via Kalman filtering and smoothing.
For our model, Markovian GPVAE (MGPVAE), we show on a variety of high-dimensional temporal tasks that our method performs favourably compared to existing approaches.
arXiv Detail & Related papers (2022-07-12T14:10:01Z) - Scaling Gaussian Process Optimization by Evaluating a Few Unique
Candidates Multiple Times [119.41129787351092]
We show that sequential black-box optimization based on GPs can be made efficient by sticking to a candidate solution for multiple evaluation steps.
We modify two well-established GP-Opt algorithms, GP-UCB and GP-EI to adapt rules from batched GP-Opt.
arXiv Detail & Related papers (2022-01-30T20:42:14Z) - Non-Gaussian Gaussian Processes for Few-Shot Regression [71.33730039795921]
We propose an invertible ODE-based mapping that operates on each component of the random variable vectors and shares the parameters across all of them.
NGGPs outperform the competing state-of-the-art approaches on a diversified set of benchmarks and applications.
arXiv Detail & Related papers (2021-10-26T10:45:25Z) - Encoding spatiotemporal priors with VAEs for small-area estimation [2.4783465852664324]
We propose a deep generative modelling approach to tackle a noveltemporal setting.
We approximate a class of prior samplings through prior fitting of a variational autoencoder (VAE)
VAE allows inference to become incredibly efficient due to independently distributed latent latent Gaussian space representation.
We demonstrate the utility of our VAE two stage approach on Bayesian, small-area estimation tasks.
arXiv Detail & Related papers (2021-10-20T08:14:15Z) - Incremental Ensemble Gaussian Processes [53.3291389385672]
We propose an incremental ensemble (IE-) GP framework, where an EGP meta-learner employs an it ensemble of GP learners, each having a unique kernel belonging to a prescribed kernel dictionary.
With each GP expert leveraging the random feature-based approximation to perform online prediction and model update with it scalability, the EGP meta-learner capitalizes on data-adaptive weights to synthesize the per-expert predictions.
The novel IE-GP is generalized to accommodate time-varying functions by modeling structured dynamics at the EGP meta-learner and within each GP learner.
arXiv Detail & Related papers (2021-10-13T15:11:25Z) - Scalable Variational Gaussian Processes via Harmonic Kernel
Decomposition [54.07797071198249]
We introduce a new scalable variational Gaussian process approximation which provides a high fidelity approximation while retaining general applicability.
We demonstrate that, on a range of regression and classification problems, our approach can exploit input space symmetries such as translations and reflections.
Notably, our approach achieves state-of-the-art results on CIFAR-10 among pure GP models.
arXiv Detail & Related papers (2021-06-10T18:17:57Z) - MuyGPs: Scalable Gaussian Process Hyperparameter Estimation Using Local
Cross-Validation [1.2233362977312945]
We present MuyGPs, a novel efficient GP hyper parameter estimation method.
MuyGPs builds upon prior methods that take advantage of the nearest neighbors structure of the data.
We show that our method outperforms all known competitors both in terms of time-to-solution and the root mean squared error of the predictions.
arXiv Detail & Related papers (2021-04-29T18:10:21Z) - Scalable Gaussian Process Variational Autoencoders [17.345687261000045]
We propose a new scalable GP-VAE model that outperforms existing approaches in terms of runtime and memory footprint, is easy to implement, and allows for joint end-to-end optimization of all components.
arXiv Detail & Related papers (2020-10-26T10:26:02Z) - Likelihood-Free Inference with Deep Gaussian Processes [70.74203794847344]
Surrogate models have been successfully used in likelihood-free inference to decrease the number of simulator evaluations.
We propose a Deep Gaussian Process (DGP) surrogate model that can handle more irregularly behaved target distributions.
Our experiments show how DGPs can outperform GPs on objective functions with multimodal distributions and maintain a comparable performance in unimodal cases.
arXiv Detail & Related papers (2020-06-18T14:24:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.