On Implicit Regularization in $\beta$-VAEs
- URL: http://arxiv.org/abs/2002.00041v4
- Date: Mon, 28 Dec 2020 22:36:36 GMT
- Title: On Implicit Regularization in $\beta$-VAEs
- Authors: Abhishek Kumar, Ben Poole
- Abstract summary: We study the regularizing effects of variational distributions on learning in generative models from two perspectives.
First, we analyze the role that the choice of variational family plays in uniqueness to the learned model by restricting the set of optimal generative models.
Second, we study the regularization effect of the variational family on the local geometry of the decoding model.
- Score: 32.674190005384204
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While the impact of variational inference (VI) on posterior inference in a
fixed generative model is well-characterized, its role in regularizing a
learned generative model when used in variational autoencoders (VAEs) is poorly
understood. We study the regularizing effects of variational distributions on
learning in generative models from two perspectives. First, we analyze the role
that the choice of variational family plays in imparting uniqueness to the
learned model by restricting the set of optimal generative models. Second, we
study the regularization effect of the variational family on the local geometry
of the decoding model. This analysis uncovers the regularizer implicit in the
$\beta$-VAE objective, and leads to an approximation consisting of a
deterministic autoencoding objective plus analytic regularizers that depend on
the Hessian or Jacobian of the decoding model, unifying VAEs with recent
heuristics proposed for training regularized autoencoders. We empirically
verify these findings, observing that the proposed deterministic objective
exhibits similar behavior to the $\beta$-VAE in terms of objective value and
sample quality.
Related papers
- Forward $\chi^2$ Divergence Based Variational Importance Sampling [2.841087763205822]
We introduce a novel variational importance sampling (VIS) approach that directly estimates and maximizes the log-likelihood.
We apply VIS to various popular latent variable models, including mixture models, variational auto-encoders, and partially observable generalized linear models.
arXiv Detail & Related papers (2023-11-04T21:46:28Z) - Mitigate Domain Shift by Primary-Auxiliary Objectives Association for
Generalizing Person ReID [39.98444065846305]
ReID models struggle in learning domain-invariant representation solely through training on an instance classification objective.
We introduce a method that guides model learning of the primary ReID instance classification objective by a concurrent auxiliary learning objective on weakly labeled pedestrian saliency detection.
Our model can be extended with the recent test-time diagram to form the PAOA+, which performs on-the-fly optimization against the auxiliary objective.
arXiv Detail & Related papers (2023-10-24T15:15:57Z) - Value-Distributional Model-Based Reinforcement Learning [59.758009422067]
Quantifying uncertainty about a policy's long-term performance is important to solve sequential decision-making tasks.
We study the problem from a model-based Bayesian reinforcement learning perspective.
We propose Epistemic Quantile-Regression (EQR), a model-based algorithm that learns a value distribution function.
arXiv Detail & Related papers (2023-08-12T14:59:19Z) - Local Disentanglement in Variational Auto-Encoders Using Jacobian $L_1$
Regularization [21.80539548847009]
Variational Auto-Encoders (VAEs) and their extensions have been shown to align latent variables with PCA directions.
We propose applying an $L_$1 loss to the VAE's generative Jacobian during training to encourage local latent variable alignment.
arXiv Detail & Related papers (2021-06-05T15:40:55Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Generative Temporal Difference Learning for Infinite-Horizon Prediction [101.59882753763888]
We introduce the $gamma$-model, a predictive model of environment dynamics with an infinite probabilistic horizon.
We discuss how its training reflects an inescapable tradeoff between training-time and testing-time compounding errors.
arXiv Detail & Related papers (2020-10-27T17:54:12Z) - Estimating Generalization under Distribution Shifts via Domain-Invariant
Representations [75.74928159249225]
We use a set of domain-invariant predictors as a proxy for the unknown, true target labels.
The error of the resulting risk estimate depends on the target risk of the proxy model.
arXiv Detail & Related papers (2020-07-06T17:21:24Z) - Decision-Making with Auto-Encoding Variational Bayes [71.44735417472043]
We show that a posterior approximation distinct from the variational distribution should be used for making decisions.
Motivated by these theoretical results, we propose learning several approximate proposals for the best model.
In addition to toy examples, we present a full-fledged case study of single-cell RNA sequencing.
arXiv Detail & Related papers (2020-02-17T19:23:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.