Embrace the Gap: VAEs Perform Independent Mechanism Analysis
- URL: http://arxiv.org/abs/2206.02416v1
- Date: Mon, 6 Jun 2022 08:19:19 GMT
- Title: Embrace the Gap: VAEs Perform Independent Mechanism Analysis
- Authors: Patrik Reizinger, Luigi Gresele, Jack Brady, Julius von K\"ugelgen,
Dominik Zietlow, Bernhard Sch\"olkopf, Georg Martius, Wieland Brendel, Michel
Besserve
- Abstract summary: We study nonlinear VAEs in the limit of near-deterministic decoders.
We show that VAEs uncover the true latent factors when the data generating process satisfies the IMA assumption.
- Score: 36.686468842036305
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Variational autoencoders (VAEs) are a popular framework for modeling complex
data distributions; they can be efficiently trained via variational inference
by maximizing the evidence lower bound (ELBO), at the expense of a gap to the
exact (log-)marginal likelihood. While VAEs are commonly used for
representation learning, it is unclear why ELBO maximization would yield useful
representations, since unregularized maximum likelihood estimation cannot
invert the data-generating process. Yet, VAEs often succeed at this task. We
seek to elucidate this apparent paradox by studying nonlinear VAEs in the limit
of near-deterministic decoders. We first prove that, in this regime, the
optimal encoder approximately inverts the decoder -- a commonly used but
unproven conjecture -- which we refer to as {\em self-consistency}. Leveraging
self-consistency, we show that the ELBO converges to a regularized
log-likelihood. This allows VAEs to perform what has recently been termed
independent mechanism analysis (IMA): it adds an inductive bias towards
decoders with column-orthogonal Jacobians, which helps recovering the true
latent factors. The gap between ELBO and log-likelihood is therefore welcome,
since it bears unanticipated benefits for nonlinear representation learning. In
experiments on synthetic and image data, we show that VAEs uncover the true
latent factors when the data generating process satisfies the IMA assumption.
Related papers
- Symmetric Equilibrium Learning of VAEs [56.56929742714685]
We view variational autoencoders (VAEs) as decoder-encoder pairs, which map distributions in the data space to distributions in the latent space and vice versa.
We propose a Nash equilibrium learning approach, which is symmetric with respect to the encoder and decoder and allows learning VAEs in situations where both the data and the latent distributions are accessible only by sampling.
arXiv Detail & Related papers (2023-07-19T10:27:34Z) - Exploiting Independent Instruments: Identification and Distribution
Generalization [3.701112941066256]
We exploit the independence for distribution generalization by taking into account higher moments.
We prove that the proposed estimator is invariant to distributional shifts on the instruments.
These results hold even in the under-identified case where the instruments are not sufficiently rich to identify the causal function.
arXiv Detail & Related papers (2022-02-03T21:49:04Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - VAE Approximation Error: ELBO and Conditional Independence [78.72292013299868]
This paper analyzes VAE approximation errors caused by the combination of the ELBO objective with the choice of the encoder probability family.
We show that the ELBO subset can not be enlarged, and the respective error cannot be decreased, by only considering deeper encoder networks.
arXiv Detail & Related papers (2021-02-18T12:54:42Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - Generalizing Variational Autoencoders with Hierarchical Empirical Bayes [6.273154057349038]
We present Hierarchical Empirical Bayes Autoencoder (HEBAE), a computationally stable framework for probabilistic generative models.
Our key contributions are two-fold. First, we make gains by placing a hierarchical prior over the encoding distribution, enabling us to adaptively balance the trade-off between minimizing the reconstruction loss function and avoiding over-regularization.
arXiv Detail & Related papers (2020-07-20T18:18:39Z) - Variational Auto-Encoder: not all failures are equal [0.0]
We show how sharpness learning addresses the notorious VAE blurriness issue.
The paper is backed upon experiments on artificial data, MNIST and CelebA, showing how sharpness learning addresses the notorious VAE blurriness issue.
arXiv Detail & Related papers (2020-03-04T09:48:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.