Training Latent Variable Models with Auto-encoding Variational Bayes: A
Tutorial
- URL: http://arxiv.org/abs/2208.07818v1
- Date: Tue, 16 Aug 2022 16:07:05 GMT
- Title: Training Latent Variable Models with Auto-encoding Variational Bayes: A
Tutorial
- Authors: Yang Zhi-Han
- Abstract summary: Auto-encoding Variational Bayes (AEVB) is a powerful and general algorithm for fitting latent variable models.
In this tutorial, we focus on motivating AEVB from the classic Expectation Maximization (EM) algorithm.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Auto-encoding Variational Bayes (AEVB) is a powerful and general algorithm
for fitting latent variable models (a promising direction for unsupervised
learning), and is well-known for training the Variational Auto-Encoder (VAE).
In this tutorial, we focus on motivating AEVB from the classic Expectation
Maximization (EM) algorithm, as opposed to from deterministic auto-encoders.
Though natural and somewhat self-evident, the connection between EM and AEVB is
not emphasized in the recent deep learning literature, and we believe that
emphasizing this connection can improve the community's understanding of AEVB.
In particular, we find it especially helpful to view (1) optimizing the
evidence lower bound (ELBO) with respect to inference parameters as approximate
E-step and (2) optimizing ELBO with respect to generative parameters as
approximate M-step; doing both simultaneously as in AEVB is then simply
tightening and pushing up ELBO at the same time. We discuss how approximate
E-step can be interpreted as performing variational inference. Important
concepts such as amortization and the reparametrization trick are discussed in
great detail. Finally, we derive from scratch the AEVB training procedures of a
non-deep and several deep latent variable models, including VAE, Conditional
VAE, Gaussian Mixture VAE and Variational RNN. It is our hope that readers
would recognize AEVB as a general algorithm that can be used to fit a wide
range of latent variable models (not just VAE), and apply AEVB to such models
that arise in their own fields of research. PyTorch code for all included
models are publicly available.
Related papers
- LAMBO: Large AI Model Empowered Edge Intelligence [71.56135386994119]
Next-generation edge intelligence is anticipated to benefit various applications via offloading techniques.
Traditional offloading architectures face several issues, including heterogeneous constraints, partial perception, uncertain generalization, and lack of tractability.
We propose a Large AI Model-Based Offloading (LAMBO) framework with over one billion parameters for solving these problems.
arXiv Detail & Related papers (2023-08-29T07:25:42Z) - Revisiting Structured Variational Autoencoders [11.998116457994994]
Structured variational autoencoders (SVAEs) combine probabilistic graphical model priors on latent variables, deep neural networks to link latent variables to observed data, and structure-exploiting algorithms for approximate posterior inference.
Despite their conceptual elegance, SVAEs have proven difficult to implement, and more general approaches have been favored in practice.
Here, we revisit SVAEs using modern machine learning tools and demonstrate their advantages over more general alternatives in terms of both accuracy and efficiency.
arXiv Detail & Related papers (2023-05-25T23:51:18Z) - AAVAE: Augmentation-Augmented Variational Autoencoders [43.73699420145321]
We introduce augmentation-augmented variational autoencoders (AAVAE), a third approach to self-supervised learning based on autoencoding.
We empirically evaluate the proposed AAVAE on image classification, similar to how recent contrastive and non-contrastive learning algorithms have been evaluated.
arXiv Detail & Related papers (2021-07-26T17:04:30Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - Unbiased Gradient Estimation for Variational Auto-Encoders using Coupled
Markov Chains [34.77971292478243]
The variational auto-encoder (VAE) is a deep latent variable model that has two neural networks in an autoencoder-like architecture.
We develop a training scheme for VAEs by introducing unbiased estimators of the log-likelihood gradient.
We show experimentally that VAEs fitted with unbiased estimators exhibit better predictive performance.
arXiv Detail & Related papers (2020-10-05T08:11:55Z) - NVAE: A Deep Hierarchical Variational Autoencoder [102.29977384039805]
We propose a deep hierarchical VAE built for image generation using depth-wise separable convolutions and batch normalization.
We show that NVAE achieves state-of-the-art results among non-autoregressive likelihood-based models.
To the best of our knowledge, NVAE is the first successful VAE applied to natural images as large as 256$times $256 pixels.
arXiv Detail & Related papers (2020-07-08T04:56:56Z) - Simple and Effective VAE Training with Calibrated Decoders [123.08908889310258]
Variational autoencoders (VAEs) provide an effective and simple method for modeling complex distributions.
We study the impact of calibrated decoders, which learn the uncertainty of the decoding distribution.
We propose a simple but novel modification to the commonly used Gaussian decoder, which computes the prediction variance analytically.
arXiv Detail & Related papers (2020-06-23T17:57:47Z) - A Convolutional Deep Markov Model for Unsupervised Speech Representation
Learning [32.59760685342343]
Probabilistic Latent Variable Models provide an alternative to self-supervised learning approaches for linguistic representation learning from speech.
In this work, we propose ConvDMM, a Gaussian state-space model with non-linear emission and transition functions modelled by deep neural networks.
When trained on a large scale speech dataset (LibriSpeech), ConvDMM produces features that significantly outperform multiple self-supervised feature extracting methods.
arXiv Detail & Related papers (2020-06-03T21:50:20Z) - On the Encoder-Decoder Incompatibility in Variational Text Modeling and
Beyond [82.18770740564642]
Variational autoencoders (VAEs) combine latent variables with amortized variational inference.
We observe the encoder-decoder incompatibility that leads to poor parameterizations of the data manifold.
We propose Coupled-VAE, which couples a VAE model with a deterministic autoencoder with the same structure.
arXiv Detail & Related papers (2020-04-20T10:34:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.