The Transitive Information Theory and its Application to Deep Generative
Models
- URL: http://arxiv.org/abs/2203.05074v1
- Date: Wed, 9 Mar 2022 22:35:02 GMT
- Title: The Transitive Information Theory and its Application to Deep Generative
Models
- Authors: Trung Ngo and Ville Hautam\"aki and Merja Hein\"aniemi
- Abstract summary: Variational Autoencoder (VAE) could be pushed in two opposite directions.
Existing methods narrow the issues to the rate-distortion trade-off between compression and reconstruction.
We develop a system that learns a hierarchy of disentangled representation together with a mechanism for recombining the learned representation for generalization.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Paradoxically, a Variational Autoencoder (VAE) could be pushed in two
opposite directions, utilizing powerful decoder model for generating realistic
images but collapsing the learned representation, or increasing regularization
coefficient for disentangling representation but ultimately generating blurry
examples. Existing methods narrow the issues to the rate-distortion trade-off
between compression and reconstruction. We argue that a good reconstruction
model does learn high capacity latents that encode more details, however, its
use is hindered by two major issues: the prior is random noise which is
completely detached from the posterior and allow no controllability in the
generation; mean-field variational inference doesn't enforce hierarchy
structure which makes the task of recombining those units into plausible novel
output infeasible. As a result, we develop a system that learns a hierarchy of
disentangled representation together with a mechanism for recombining the
learned representation for generalization. This is achieved by introducing a
minimal amount of inductive bias to learn controllable prior for the VAE. The
idea is supported by here developed transitive information theory, that is, the
mutual information between two target variables could alternately be maximized
through the mutual information to the third variable, thus bypassing the
rate-distortion bottleneck in VAE design. In particular, we show that our
model, named SemafoVAE (inspired by the similar concept in computer science),
could generate high-quality examples in a controllable manner, perform smooth
traversals of the disentangled factors and intervention at a different level of
representation hierarchy.
Related papers
- Disentanglement with Factor Quantized Variational Autoencoders [11.086500036180222]
We propose a discrete variational autoencoder (VAE) based model where the ground truth information about the generative factors are not provided to the model.
We demonstrate the advantages of learning discrete representations over learning continuous representations in facilitating disentanglement.
Our method called FactorQVAE is the first method that combines optimization based disentanglement approaches with discrete representation learning.
arXiv Detail & Related papers (2024-09-23T09:33:53Z) - DiffEnc: Variational Diffusion with a Learned Encoder [14.045374947755922]
We introduce a data- and depth-dependent mean function in the diffusion process, which leads to a modified diffusion loss.
Our proposed framework, DiffEnc, achieves a statistically significant improvement in likelihood on CIFAR-10.
arXiv Detail & Related papers (2023-10-30T17:54:36Z) - How to train your VAE [0.0]
Variational Autoencoders (VAEs) have become a cornerstone in generative modeling and representation learning within machine learning.
This paper explores interpreting the Kullback-Leibler (KL) Divergence, a critical component within the Evidence Lower Bound (ELBO)
The proposed method redefines the ELBO with a mixture of Gaussians for the posterior probability, introduces a regularization term, and employs a PatchGAN discriminator to enhance texture realism.
arXiv Detail & Related papers (2023-09-22T19:52:28Z) - Enhancing Multiple Reliability Measures via Nuisance-extended
Information Bottleneck [77.37409441129995]
In practical scenarios where training data is limited, many predictive signals in the data can be rather from some biases in data acquisition.
We consider an adversarial threat model under a mutual information constraint to cover a wider class of perturbations in training.
We propose an autoencoder-based training to implement the objective, as well as practical encoder designs to facilitate the proposed hybrid discriminative-generative training.
arXiv Detail & Related papers (2023-03-24T16:03:21Z) - DIFFormer: Scalable (Graph) Transformers Induced by Energy Constrained
Diffusion [66.21290235237808]
We introduce an energy constrained diffusion model which encodes a batch of instances from a dataset into evolutionary states.
We provide rigorous theory that implies closed-form optimal estimates for the pairwise diffusion strength among arbitrary instance pairs.
Experiments highlight the wide applicability of our model as a general-purpose encoder backbone with superior performance in various tasks.
arXiv Detail & Related papers (2023-01-23T15:18:54Z) - Adaptive Discrete Communication Bottlenecks with Dynamic Vector
Quantization [76.68866368409216]
We propose learning to dynamically select discretization tightness conditioned on inputs.
We show that dynamically varying tightness in communication bottlenecks can improve model performance on visual reasoning and reinforcement learning tasks.
arXiv Detail & Related papers (2022-02-02T23:54:26Z) - Towards Robust and Adaptive Motion Forecasting: A Causal Representation
Perspective [72.55093886515824]
We introduce a causal formalism of motion forecasting, which casts the problem as a dynamic process with three groups of latent variables.
We devise a modular architecture that factorizes the representations of invariant mechanisms and style confounders to approximate a causal graph.
Experiment results on synthetic and real datasets show that our three proposed components significantly improve the robustness and reusability of the learned motion representations.
arXiv Detail & Related papers (2021-11-29T18:59:09Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - InfoVAEGAN : learning joint interpretable representations by information
maximization and maximum likelihood [15.350366047108103]
We propose a representation learning algorithm which combines the inference abilities of Variational Autoencoders (VAE) with the capability of Generative Adversarial Networks (GAN)
The proposed model, called InfoVAEGAN, consists of three networks:generative Generator and Discriminator.
arXiv Detail & Related papers (2021-07-09T22:38:10Z) - High-Fidelity Synthesis with Disentangled Representation [60.19657080953252]
We propose an Information-Distillation Generative Adrial Network (ID-GAN) for disentanglement learning and high-fidelity synthesis.
Our method learns disentangled representation using VAE-based models, and distills the learned representation with an additional nuisance variable to the separate GAN-based generator for high-fidelity synthesis.
Despite the simplicity, we show that the proposed method is highly effective, achieving comparable image generation quality to the state-of-the-art methods using the disentangled representation.
arXiv Detail & Related papers (2020-01-13T14:39:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.