Bayesian Non-linear Latent Variable Modeling via Random Fourier Features
- URL: http://arxiv.org/abs/2306.08352v1
- Date: Wed, 14 Jun 2023 08:42:10 GMT
- Title: Bayesian Non-linear Latent Variable Modeling via Random Fourier Features
- Authors: Michael Minyi Zhang, Gregory W. Gundersen, Barbara E. Engelhardt
- Abstract summary: We present a method to perform Markov chain Monte Carlo inference for generalized nonlinear latent variable modeling.
Inference forVMs is computationally tractable only when the data likelihood is Gaussian.
We show that we can generalizeVMs to non-Gaussian observations, such as Poisson, negative binomial, and multinomial distributions.
- Score: 7.856578780790166
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Gaussian process latent variable model (GPLVM) is a popular probabilistic
method used for nonlinear dimension reduction, matrix factorization, and
state-space modeling. Inference for GPLVMs is computationally tractable only
when the data likelihood is Gaussian. Moreover, inference for GPLVMs has
typically been restricted to obtaining maximum a posteriori point estimates,
which can lead to overfitting, or variational approximations, which
mischaracterize the posterior uncertainty. Here, we present a method to perform
Markov chain Monte Carlo (MCMC) inference for generalized Bayesian nonlinear
latent variable modeling. The crucial insight necessary to generalize GPLVMs to
arbitrary observation models is that we approximate the kernel function in the
Gaussian process mappings with random Fourier features; this allows us to
compute the gradient of the posterior in closed form with respect to the latent
variables. We show that we can generalize GPLVMs to non-Gaussian observations,
such as Poisson, negative binomial, and multinomial distributions, using our
random feature latent variable model (RFLVM). Our generalized RFLVMs perform on
par with state-of-the-art latent variable models on a wide range of
applications, including motion capture, images, and text data for the purpose
of estimating the latent structure and imputing the missing data of these
complex data sets.
Related papers
- von Mises Quasi-Processes for Bayesian Circular Regression [57.88921637944379]
We explore a family of expressive and interpretable distributions over circle-valued random functions.
The resulting probability model has connections with continuous spin models in statistical physics.
For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Markov Chain Monte Carlo sampling.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - Scaling and renormalization in high-dimensional regression [72.59731158970894]
This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models.
We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning.
arXiv Detail & Related papers (2024-05-01T15:59:00Z) - Preventing Model Collapse in Gaussian Process Latent Variable Models [11.45681373843122]
This paper theoretically examines the impact of projection variance on model collapse through the lens of a linear FourierVM.
We tackle model collapse due to inadequate kernel flexibility by integrating the spectral mixture (SM) kernel and a differentiable random feature (RFF) kernel approximation.
The proposedVM, named advisedRFLVM, is evaluated across diverse datasets and consistently outperforms various competing models.
arXiv Detail & Related papers (2024-04-02T06:58:41Z) - Fusion of Gaussian Processes Predictions with Monte Carlo Sampling [61.31380086717422]
In science and engineering, we often work with models designed for accurate prediction of variables of interest.
Recognizing that these models are approximations of reality, it becomes desirable to apply multiple models to the same data and integrate their outcomes.
arXiv Detail & Related papers (2024-03-03T04:21:21Z) - Diffusion models for probabilistic programming [56.47577824219207]
Diffusion Model Variational Inference (DMVI) is a novel method for automated approximate inference in probabilistic programming languages (PPLs)
DMVI is easy to implement, allows hassle-free inference in PPLs without the drawbacks of, e.g., variational inference using normalizing flows, and does not make any constraints on the underlying neural network model.
arXiv Detail & Related papers (2023-11-01T12:17:05Z) - Probabilistic Unrolling: Scalable, Inverse-Free Maximum Likelihood
Estimation for Latent Gaussian Models [69.22568644711113]
We introduce probabilistic unrolling, a method that combines Monte Carlo sampling with iterative linear solvers to circumvent matrix inversions.
Our theoretical analyses reveal that unrolling and backpropagation through the iterations of the solver can accelerate gradient estimation for maximum likelihood estimation.
In experiments on simulated and real data, we demonstrate that probabilistic unrolling learns latent Gaussian models up to an order of magnitude faster than gradient EM, with minimal losses in model performance.
arXiv Detail & Related papers (2023-06-05T21:08:34Z) - Generalised Gaussian Process Latent Variable Models (GPLVM) with
Stochastic Variational Inference [9.468270453795409]
We study the doubly formulation of the BayesianVM model amenable with minibatch training.
We show how this framework is compatible with different latent variable formulations and perform experiments to compare a suite of models.
We demonstrate how we can train in the presence of massively missing data and obtain high-fidelity reconstructions.
arXiv Detail & Related papers (2022-02-25T21:21:51Z) - Inverting brain grey matter models with likelihood-free inference: a
tool for trustable cytoarchitecture measurements [62.997667081978825]
characterisation of the brain grey matter cytoarchitecture with quantitative sensitivity to soma density and volume remains an unsolved challenge in dMRI.
We propose a new forward model, specifically a new system of equations, requiring a few relatively sparse b-shells.
We then apply modern tools from Bayesian analysis known as likelihood-free inference (LFI) to invert our proposed model.
arXiv Detail & Related papers (2021-11-15T09:08:27Z) - Learning Nonparametric Volterra Kernels with Gaussian Processes [0.0]
This paper introduces a method for the nonparametric Bayesian learning of nonlinear operators, through the use of the Volterra series with kernels represented using Gaussian processes (GPs)
When the input function to the operator is unobserved and has a GP prior, the NVKM constitutes a powerful method for both single and multiple output regression, and can be viewed as a nonlinear and nonparametric latent force model.
arXiv Detail & Related papers (2021-06-10T08:21:00Z) - Latent variable modeling with random features [7.856578780790166]
We develop a family of nonlinear dimension reduction models that are easily to non-Gaussian data likelihoods.
Our generalized RFLVMs produce results comparable with other state-of-the-art dimension reduction methods.
arXiv Detail & Related papers (2020-06-19T14:12:05Z) - tvGP-VAE: Tensor-variate Gaussian Process Prior Variational Autoencoder [0.0]
tvGP-VAE is able to explicitly model correlation via the use of kernel functions.
We show that the choice of which correlation structures to explicitly represent in the latent space has a significant impact on model performance.
arXiv Detail & Related papers (2020-06-08T17:59:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.