Provably Reliable Large-Scale Sampling from Gaussian Processes
- URL: http://arxiv.org/abs/2211.08036v1
- Date: Tue, 15 Nov 2022 10:36:21 GMT
- Title: Provably Reliable Large-Scale Sampling from Gaussian Processes
- Authors: Anthony Stephenson, Robert Allison, Edward Pyzer-Knapp
- Abstract summary: We demonstrate how to generate a size (n) sample from approximate Gaussian process (GP) models.
We provide guarantees that, with high probability, the sample is indistinguishable from a sample from the desired GP.
- Score: 3.6417668958891793
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: When comparing approximate Gaussian process (GP) models, it can be helpful to
be able to generate data from any GP. If we are interested in how approximate
methods perform at scale, we may wish to generate very large synthetic datasets
to evaluate them. Na\"{i}vely doing so would cost \(\mathcal{O}(n^3)\) flops
and \(\mathcal{O}(n^2)\) memory to generate a size \(n\) sample. We demonstrate
how to scale such data generation to large \(n\) whilst still providing
guarantees that, with high probability, the sample is indistinguishable from a
sample from the desired GP.
Related papers
- Gaussian Processes Sampling with Sparse Grids under Additive Schwarz Preconditioner [6.408773096179187]
We propose a scalable algorithm for sampling random realizations of the prior and posterior of GP models.
The proposed algorithm leverages inducing points approximation with sparse grids, as well as additive Schwarz preconditioners.
arXiv Detail & Related papers (2024-08-01T00:19:36Z) - A Good Score Does not Lead to A Good Generative Model [14.752242187781107]
Score-based Generative Models (SGMs) is one leading method in generative modeling.
We show that SGMs can generate samples from a distribution that is close to the ground-truth if the underlying score function is learned well.
arXiv Detail & Related papers (2024-01-10T00:17:36Z) - Implicit Manifold Gaussian Process Regression [49.0787777751317]
Gaussian process regression is widely used to provide well-calibrated uncertainty estimates.
It struggles with high-dimensional data because of the implicit low-dimensional manifold upon which the data actually lies.
In this paper we propose a technique capable of inferring implicit structure directly from data (labeled and unlabeled) in a fully differentiable way.
arXiv Detail & Related papers (2023-10-30T09:52:48Z) - Adaptive Sketches for Robust Regression with Importance Sampling [64.75899469557272]
We introduce data structures for solving robust regression through gradient descent (SGD)
Our algorithm effectively runs $T$ steps of SGD with importance sampling while using sublinear space and just making a single pass over the data.
arXiv Detail & Related papers (2022-07-16T03:09:30Z) - Algorithmic Gaussianization through Sketching: Converting Data into
Sub-gaussian Random Designs [22.925108493465363]
We provide an algorithmic framework for gaussianizing data distributions via averaging.
We show that it is possible to efficiently construct data sketches that are nearly indistinguishable from sub-gaussian random designs.
arXiv Detail & Related papers (2022-06-21T12:16:45Z) - Convergence for score-based generative modeling with polynomial
complexity [9.953088581242845]
We prove the first convergence guarantees for the core mechanic behind Score-based generative modeling.
Compared to previous works, we do not incur error that grows exponentially in time or that suffers from a curse of dimensionality.
We show that a predictor-corrector gives better convergence than using either portion alone.
arXiv Detail & Related papers (2022-06-13T14:57:35Z) - Probabilistic Registration for Gaussian Process 3D shape modelling in
the presence of extensive missing data [63.8376359764052]
We propose a shape fitting/registration method based on a Gaussian Processes formulation, suitable for shapes with extensive regions of missing data.
Experiments are conducted both for a 2D small dataset with diverse transformations and a 3D dataset of ears.
arXiv Detail & Related papers (2022-03-26T16:48:27Z) - Oops I Took A Gradient: Scalable Sampling for Discrete Distributions [53.3142984019796]
We show that this approach outperforms generic samplers in a number of difficult settings.
We also demonstrate the use of our improved sampler for training deep energy-based models on high dimensional discrete data.
arXiv Detail & Related papers (2021-02-08T20:08:50Z) - Revisiting the Sample Complexity of Sparse Spectrum Approximation of
Gaussian Processes [60.479499225746295]
We introduce a new scalable approximation for Gaussian processes with provable guarantees which hold simultaneously over its entire parameter space.
Our approximation is obtained from an improved sample complexity analysis for sparse spectrum Gaussian processes (SSGPs)
arXiv Detail & Related papers (2020-11-17T05:41:50Z) - Linear-time inference for Gaussian Processes on one dimension [17.77516394591124]
We investigate data sampled on one dimension for which state-space models are popular due to their linearly-scaling computational costs.
We provide the first general proof of conjecture that state-space models are general, able to approximate any one-dimensional Gaussian Processes.
We develop parallelized algorithms for performing inference and learning in the LEG model, test the algorithm on real and synthetic data, and demonstrate scaling to datasets with billions of samples.
arXiv Detail & Related papers (2020-03-11T23:20:13Z) - Efficiently Sampling Functions from Gaussian Process Posteriors [76.94808614373609]
We propose an easy-to-use and general-purpose approach for fast posterior sampling.
We demonstrate how decoupled sample paths accurately represent Gaussian process posteriors at a fraction of the usual cost.
arXiv Detail & Related papers (2020-02-21T14:03:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.