Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness
- URL: http://arxiv.org/abs/2110.12381v1
- Date: Sun, 24 Oct 2021 07:58:13 GMT
- Title: Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness
- Authors: Dazhong Shen and Chuan Qin and Chao Wang and Hengshu Zhu and Enhong
Chen and Hui Xiong
- Abstract summary: Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
- Score: 61.827054365139645
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As one of the most popular generative models, Variational Autoencoder (VAE)
approximates the posterior of latent variables based on amortized variational
inference. However, when the decoder network is sufficiently expressive, VAE
may lead to posterior collapse; that is, uninformative latent representations
may be learned. To this end, in this paper, we propose an alternative model,
DU-VAE, for learning a more Diverse and less Uncertain latent space, and thus
the representation can be learned in a meaningful and compact manner.
Specifically, we first theoretically demonstrate that it will result in better
latent space with high diversity and low uncertainty awareness by controlling
the distribution of posterior's parameters across the whole data accordingly.
Then, without the introduction of new loss terms or modifying training
strategies, we propose to exploit Dropout on the variances and
Batch-Normalization on the means simultaneously to regularize their
distributions implicitly. Furthermore, to evaluate the generalization effect,
we also exploit DU-VAE for inverse autoregressive flow based-VAE (VAE-IAF)
empirically. Finally, extensive experiments on three benchmark datasets clearly
show that our approach can outperform state-of-the-art baselines on both
likelihood estimation and underlying classification tasks.
Related papers
- Protect Before Generate: Error Correcting Codes within Discrete Deep Generative Models [3.053842954605396]
We introduce a novel method that enhances variational inference in discrete latent variable models.
We leverage Error Correcting Codes (ECCs) to introduce redundancy in the latent representations.
This redundancy is then exploited by the variational posterior to yield more accurate estimates.
arXiv Detail & Related papers (2024-10-10T11:59:58Z) - Singular Value Penalization and Semantic Data Augmentation for Fully
Test-Time Adaptation [5.891527229524256]
Test-time adaptation (FTTA) adapts a model that is trained on a source domain to a target domain during the testing phase.
We propose maximizing the sum of singular values while minimizing their variance.
This enables the model's focus toward the smaller singular values, enhancing discriminability between more challenging classes and effectively increasing the diversity of prediction results.
arXiv Detail & Related papers (2023-12-10T01:08:56Z) - Variational Distillation for Multi-View Learning [104.17551354374821]
We design several variational information bottlenecks to exploit two key characteristics for multi-view representation learning.
Under rigorously theoretical guarantee, our approach enables IB to grasp the intrinsic correlation between observations and semantic labels.
arXiv Detail & Related papers (2022-06-20T03:09:46Z) - Covariate-informed Representation Learning with Samplewise Optimal
Identifiable Variational Autoencoders [15.254297587065595]
Recently proposed identifiable variational autoencoder (iVAE) provides a promising approach for learning latent independent components of the data.
We develop a new approach, co-informed identifiable VAE (CI-iVAE)
In doing so, the objective function enforces the inverse relation, and learned representation contains more information of observations.
arXiv Detail & Related papers (2022-02-09T00:18:33Z) - Revisiting Consistency Regularization for Semi-Supervised Learning [80.28461584135967]
We propose an improved consistency regularization framework by a simple yet effective technique, FeatDistLoss.
Experimental results show that our model defines a new state of the art for various datasets and settings.
arXiv Detail & Related papers (2021-12-10T20:46:13Z) - InteL-VAEs: Adding Inductive Biases to Variational Auto-Encoders via
Intermediary Latents [60.785317191131284]
We introduce a simple and effective method for learning VAEs with controllable biases by using an intermediary set of latent variables.
In particular, it allows us to impose desired properties like sparsity or clustering on learned representations.
We show that this, in turn, allows InteL-VAEs to learn both better generative models and representations.
arXiv Detail & Related papers (2021-06-25T16:34:05Z) - Squared $\ell_2$ Norm as Consistency Loss for Leveraging Augmented Data
to Learn Robust and Invariant Representations [76.85274970052762]
Regularizing distance between embeddings/representations of original samples and augmented counterparts is a popular technique for improving robustness of neural networks.
In this paper, we explore these various regularization choices, seeking to provide a general understanding of how we should regularize the embeddings.
We show that the generic approach we identified (squared $ell$ regularized augmentation) outperforms several recent methods, which are each specially designed for one task.
arXiv Detail & Related papers (2020-11-25T22:40:09Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z) - Automatic Differentiation Variational Inference with Mixtures [4.995383193706478]
We show how stratified sampling may be used to enable mixture distributions as the approximate posterior.
We derive a new lower bound on the evidence analogous to the importance weighted autoencoder (IWAE)
arXiv Detail & Related papers (2020-03-03T18:12:42Z) - NestedVAE: Isolating Common Factors via Weak Supervision [45.366986365879505]
We identify the connection between the task of bias reduction and that of isolating factors common between domains.
To isolate the common factors we combine the theory of deep latent variable models with information bottleneck theory.
Two outer VAEs with shared weights attempt to reconstruct the input and infer a latent space, whilst a nested VAE attempts to reconstruct the latent representation of one image, from the latent representation of its paired image.
arXiv Detail & Related papers (2020-02-26T15:49:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.