Sparse Infinite Random Feature Latent Variable Modeling
- URL: http://arxiv.org/abs/2205.09909v1
- Date: Fri, 20 May 2022 00:29:28 GMT
- Title: Sparse Infinite Random Feature Latent Variable Modeling
- Authors: Michael Minyi Zhang
- Abstract summary: A posteriori, the number of instantiated dimensions in the latent space is guaranteed to be finite.
We show that we can obtain superior test set performance compared to previous latent variable models.
- Score: 6.063419970703021
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a non-linear, Bayesian non-parametric latent variable model where
the latent space is assumed to be sparse and infinite dimensional a priori
using an Indian buffet process prior. A posteriori, the number of instantiated
dimensions in the latent space is guaranteed to be finite. The purpose of
placing the Indian buffet process on the latent variables is to: 1.)
Automatically and probabilistically select the number of latent dimensions. 2.)
Impose sparsity in the latent space, where the Indian buffet process will
select which elements are exactly zero. Our proposed model allows for sparse,
non-linear latent variable modeling where the number of latent dimensions is
selected automatically. Inference is made tractable using the random Fourier
approximation and we can easily implement posterior inference through Markov
chain Monte Carlo sampling. This approach is amenable to many observation
models beyond the Gaussian setting. We demonstrate the utility of our method on
a variety of synthetic, biological and text datasets and show that we can
obtain superior test set performance compared to previous latent variable
models.
Related papers
- A model-free feature selection technique of feature screening and random
forest based recursive feature elimination [0.0]
We propose a model-free feature selection method for ultra-high dimensional data with mass features.
We show that the proposed method is selection consistent and $L$ consistent under weak regularity conditions.
arXiv Detail & Related papers (2023-02-15T03:39:16Z) - GFlowNet-EM for learning compositional latent variable models [115.96660869630227]
A key tradeoff in modeling the posteriors over latents is between expressivity and tractable optimization.
We propose the use of GFlowNets, algorithms for sampling from an unnormalized density.
By training GFlowNets to sample from the posterior over latents, we take advantage of their strengths as amortized variational algorithms.
arXiv Detail & Related papers (2023-02-13T18:24:21Z) - Posterior Collapse and Latent Variable Non-identifiability [54.842098835445]
We propose a class of latent-identifiable variational autoencoders, deep generative models which enforce identifiability without sacrificing flexibility.
Across synthetic and real datasets, latent-identifiable variational autoencoders outperform existing methods in mitigating posterior collapse and providing meaningful representations of the data.
arXiv Detail & Related papers (2023-01-02T06:16:56Z) - FONDUE: an algorithm to find the optimal dimensionality of the latent
representations of variational autoencoders [2.969705152497174]
In this paper, we explore the intrinsic dimension estimation (IDE) of the data and latent representations learned by VAEs.
We show that the discrepancies between theIDE of the mean and sampled representations of a VAE after only a few steps of training reveal the presence of passive variables in the latent space.
We propose FONDUE: an algorithm which quickly finds the number of latent dimensions after which the mean and sampled representations start to diverge.
arXiv Detail & Related papers (2022-09-26T15:59:54Z) - Learning and Inference in Sparse Coding Models with Langevin Dynamics [3.0600309122672726]
We describe a system capable of inference and learning in a probabilistic latent variable model.
We demonstrate this idea for a sparse coding model by deriving a continuous-time equation for inferring its latent variables via Langevin dynamics.
We show that Langevin dynamics lead to an efficient procedure for sampling from the posterior distribution in the 'L0 sparse' regime, where latent variables are encouraged to be set to zero as opposed to having a small L1 norm.
arXiv Detail & Related papers (2022-04-23T23:16:47Z) - Oops I Took A Gradient: Scalable Sampling for Discrete Distributions [53.3142984019796]
We show that this approach outperforms generic samplers in a number of difficult settings.
We also demonstrate the use of our improved sampler for training deep energy-based models on high dimensional discrete data.
arXiv Detail & Related papers (2021-02-08T20:08:50Z) - Dimension-robust Function Space MCMC With Neural Network Priors [0.0]
This paper introduces a new prior on functions spaces which scales more favourably in the dimension of the function's domain.
We show that our resulting posterior of the unknown function is amenable to sampling using Hilbert space Markov chain Monte Carlo methods.
We show that our priors are competitive and have distinct advantages over other function space priors.
arXiv Detail & Related papers (2020-12-20T14:52:57Z) - Uncertainty Inspired RGB-D Saliency Detection [70.50583438784571]
We propose the first framework to employ uncertainty for RGB-D saliency detection by learning from the data labeling process.
Inspired by the saliency data labeling process, we propose a generative architecture to achieve probabilistic RGB-D saliency detection.
Results on six challenging RGB-D benchmark datasets show our approach's superior performance in learning the distribution of saliency maps.
arXiv Detail & Related papers (2020-09-07T13:01:45Z) - Slice Sampling for General Completely Random Measures [74.24975039689893]
We present a novel Markov chain Monte Carlo algorithm for posterior inference that adaptively sets the truncation level using auxiliary slice variables.
The efficacy of the proposed algorithm is evaluated on several popular nonparametric models.
arXiv Detail & Related papers (2020-06-24T17:53:53Z) - SUMO: Unbiased Estimation of Log Marginal Probability for Latent
Variable Models [80.22609163316459]
We introduce an unbiased estimator of the log marginal likelihood and its gradients for latent variable models based on randomized truncation of infinite series.
We show that models trained using our estimator give better test-set likelihoods than a standard importance-sampling based approach for the same average computational cost.
arXiv Detail & Related papers (2020-04-01T11:49:30Z) - Generalized Gumbel-Softmax Gradient Estimator for Various Discrete
Random Variables [16.643346012854156]
Esting the gradients of nodes is one of the crucial research questions in the deep generative modeling community.
This paper proposes a general version of the Gumbel-Softmax estimator with continuous relaxation.
arXiv Detail & Related papers (2020-03-04T01:13:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.