DAVA: Disentangling Adversarial Variational Autoencoder
- URL: http://arxiv.org/abs/2303.01384v1
- Date: Thu, 2 Mar 2023 16:08:23 GMT
- Title: DAVA: Disentangling Adversarial Variational Autoencoder
- Authors: Benjamin Estermann and Roger Wattenhofer
- Abstract summary: We introduce DAVA, a novel training procedure for variational auto-encoders.
We demonstrate the ability of PIPE to positively predict the performance of downstream models in abstract reasoning.
- Score: 12.513372993000914
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The use of well-disentangled representations offers many advantages for
downstream tasks, e.g. an increased sample efficiency, or better
interpretability. However, the quality of disentangled interpretations is often
highly dependent on the choice of dataset-specific hyperparameters, in
particular the regularization strength. To address this issue, we introduce
DAVA, a novel training procedure for variational auto-encoders. DAVA completely
alleviates the problem of hyperparameter selection. We compare DAVA to models
with optimal hyperparameters. Without any hyperparameter tuning, DAVA is
competitive on a diverse range of commonly used datasets. Underlying DAVA, we
discover a necessary condition for unsupervised disentanglement, which we call
PIPE. We demonstrate the ability of PIPE to positively predict the performance
of downstream models in abstract reasoning. We also thoroughly investigate
correlations with existing supervised and unsupervised metrics. The code is
available at https://github.com/besterma/dava.
Related papers
- HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - RENs: Relevance Encoding Networks [0.0]
This paper proposes relevance encoding networks (RENs): a novel probabilistic VAE-based framework that uses the automatic relevance determination (ARD) prior in the latent space to learn the data-specific bottleneck dimensionality.
We show that the proposed model learns the relevant latent bottleneck dimensionality without compromising the representation and generation quality of the samples.
arXiv Detail & Related papers (2022-05-25T21:53:48Z) - Discrete Auto-regressive Variational Attention Models for Text Modeling [53.38382932162732]
Variational autoencoders (VAEs) have been widely applied for text modeling.
They are troubled by two challenges: information underrepresentation and posterior collapse.
We propose Discrete Auto-regressive Variational Attention Model (DAVAM) to address the challenges.
arXiv Detail & Related papers (2021-06-16T06:36:26Z) - Model Selection for Bayesian Autoencoders [25.619565817793422]
We propose to optimize the distributional sliced-Wasserstein distance between the output of the autoencoder and the empirical data distribution.
We turn our BAE into a generative model by fitting a flexible Dirichlet mixture model in the latent space.
We evaluate our approach qualitatively and quantitatively using a vast experimental campaign on a number of unsupervised learning tasks and show that, in small-data regimes where priors matter, our approach provides state-of-the-art results.
arXiv Detail & Related papers (2021-06-11T08:55:00Z) - Understanding Overparameterization in Generative Adversarial Networks [56.57403335510056]
Generative Adversarial Networks (GANs) are used to train non- concave mini-max optimization problems.
A theory has shown the importance of the gradient descent (GD) to globally optimal solutions.
We show that in an overized GAN with a $1$-layer neural network generator and a linear discriminator, the GDA converges to a global saddle point of the underlying non- concave min-max problem.
arXiv Detail & Related papers (2021-04-12T16:23:37Z) - Adversarial and Contrastive Variational Autoencoder for Sequential
Recommendation [25.37244686572865]
We propose a novel method called Adversarial and Contrastive Variational Autoencoder (ACVAE) for sequential recommendation.
We first introduce the adversarial training for sequence generation under the Adversarial Variational Bayes framework, which enables our model to generate high-quality latent variables.
Besides, when encoding the sequence, we apply a recurrent and convolutional structure to capture global and local relationships in the sequence.
arXiv Detail & Related papers (2021-03-19T09:01:14Z) - Preventing Posterior Collapse Induced by Oversmoothing in Gaussian VAE [7.845959449872641]
We propose AR-ELBO, which controls the smoothness of the model by adapting the variance parameter.
In addition, we extend VAE with alternative parameterizations on the variance parameter to deal with non-uniform or conditional data variance.
The proposed VAE extensions trained with AR-ELBO show improved Fr'echet inception distance (FID) on images generated from the MNIST and CelebA datasets.
arXiv Detail & Related papers (2021-02-17T10:00:49Z) - Simple and Effective VAE Training with Calibrated Decoders [123.08908889310258]
Variational autoencoders (VAEs) provide an effective and simple method for modeling complex distributions.
We study the impact of calibrated decoders, which learn the uncertainty of the decoding distribution.
We propose a simple but novel modification to the commonly used Gaussian decoder, which computes the prediction variance analytically.
arXiv Detail & Related papers (2020-06-23T17:57:47Z) - SUOD: Accelerating Large-Scale Unsupervised Heterogeneous Outlier
Detection [63.253850875265115]
Outlier detection (OD) is a key machine learning (ML) task for identifying abnormal objects from general samples.
We propose a modular acceleration system, called SUOD, to address it.
arXiv Detail & Related papers (2020-03-11T00:22:50Z) - Rethinking the Hyperparameters for Fine-tuning [78.15505286781293]
Fine-tuning from pre-trained ImageNet models has become the de-facto standard for various computer vision tasks.
Current practices for fine-tuning typically involve selecting an ad-hoc choice of hyper parameters.
This paper re-examines several common practices of setting hyper parameters for fine-tuning.
arXiv Detail & Related papers (2020-02-19T18:59:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.