Inference-InfoGAN: Inference Independence via Embedding Orthogonal Basis
Expansion
- URL: http://arxiv.org/abs/2110.00788v1
- Date: Sat, 2 Oct 2021 11:54:23 GMT
- Title: Inference-InfoGAN: Inference Independence via Embedding Orthogonal Basis
Expansion
- Authors: Hongxiang Jiang, Jihao Yin, Xiaoyan Luo, Fuxiang Wang
- Abstract summary: Disentanglement learning aims to construct independent and interpretable latent variables in which generative models are a popular strategy.
We propose a novel GAN-based disentanglement framework via embedding Orthogonal Basis Expansion (OBE) into InfoGAN network.
Our Inference-InfoGAN achieves higher disentanglement score in terms of FactorVAE, Separated ferenceAttribute Predictability (SAP), Mutual Information Gap (MIG) and Variation Predictability (VP) metrics without model fine-tuning.
- Score: 2.198430261120653
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Disentanglement learning aims to construct independent and interpretable
latent variables in which generative models are a popular strategy. InfoGAN is
a classic method via maximizing Mutual Information (MI) to obtain interpretable
latent variables mapped to the target space. However, it did not emphasize
independent characteristic. To explicitly infer latent variables with
inter-independence, we propose a novel GAN-based disentanglement framework via
embedding Orthogonal Basis Expansion (OBE) into InfoGAN network
(Inference-InfoGAN) in an unsupervised way. Under the OBE module, one set of
orthogonal basis can be adaptively found to expand arbitrary data with
independence property. To ensure the target-wise interpretable representation,
we add a consistence constraint between the expansion coefficients and latent
variables on the base of MI maximization. Additionally, we design an
alternating optimization step on the consistence constraint and orthogonal
requirement updating, so that the training of Inference-InfoGAN can be more
convenient. Finally, experiments validate that our proposed OBE module obtains
adaptive orthogonal basis, which can express better independent characteristics
than fixed basis expression of Discrete Cosine Transform (DCT). To depict the
performance in downstream tasks, we compared with the state-of-the-art
GAN-based and even VAE-based approaches on different datasets. Our
Inference-InfoGAN achieves higher disentanglement score in terms of FactorVAE,
Separated Attribute Predictability (SAP), Mutual Information Gap (MIG) and
Variation Predictability (VP) metrics without model fine-tuning. All the
experimental results illustrate that our method has inter-independence
inference ability because of the OBE module, and provides a good trade-off
between it and target-wise interpretability of latent variables via jointing
the alternating optimization.
Related papers
- Uniform Transformation: Refining Latent Representation in Variational Autoencoders [7.4316292428754105]
We introduce a novel adaptable three-stage Uniform Transformation (UT) module to address irregular latent distributions.
By reconfiguring irregular distributions into a uniform distribution in the latent space, our approach significantly enhances the disentanglement and interpretability of latent representations.
Empirical evaluations demonstrated the efficacy of our proposed UT module in improving disentanglement metrics across benchmark datasets.
arXiv Detail & Related papers (2024-07-02T21:46:23Z) - Disentanglement via Latent Quantization [60.37109712033694]
In this work, we construct an inductive bias towards encoding to and decoding from an organized latent space.
We demonstrate the broad applicability of this approach by adding it to both basic data-re (vanilla autoencoder) and latent-reconstructing (InfoGAN) generative models.
arXiv Detail & Related papers (2023-05-28T06:30:29Z) - Elastic Interaction Energy-Based Generative Model: Approximation in
Feature Space [14.783344918500813]
We propose a novel approach to generative modeling using a loss function based on elastic interaction energy (EIE)
The utilization of the EIE-based metric presents several advantages, including its long range property that enables consideration of global information in the distribution.
Experimental results on popular datasets, such as MNIST, FashionMNIST, CIFAR-10, and CelebA, demonstrate that our EIEG GAN model can mitigate mode collapse, enhance stability, and improve model performance.
arXiv Detail & Related papers (2023-03-19T03:39:31Z) - Variable Importance Matching for Causal Inference [73.25504313552516]
We describe a general framework called Model-to-Match that achieves these goals.
Model-to-Match uses variable importance measurements to construct a distance metric.
We operationalize the Model-to-Match framework with LASSO.
arXiv Detail & Related papers (2023-02-23T00:43:03Z) - Attitudes and Latent Class Choice Models using Machine learning [0.0]
We present a method of efficiently incorporating attitudinal indicators in the specification of Latent Class Choice Models (LCCM)
This formulation overcomes structural equations in its capability of exploring the relationship between the attitudinal indicators and the decision choice.
We test our proposed framework for estimating a Car-Sharing (CS) service subscription choice with stated preference data from Copenhagen, Denmark.
arXiv Detail & Related papers (2023-02-20T10:03:01Z) - Latent Variable Representation for Reinforcement Learning [131.03944557979725]
It remains unclear theoretically and empirically how latent variable models may facilitate learning, planning, and exploration to improve the sample efficiency of model-based reinforcement learning.
We provide a representation view of the latent variable models for state-action value functions, which allows both tractable variational learning algorithm and effective implementation of the optimism/pessimism principle.
In particular, we propose a computationally efficient planning algorithm with UCB exploration by incorporating kernel embeddings of latent variable models.
arXiv Detail & Related papers (2022-12-17T00:26:31Z) - Break The Spell Of Total Correlation In betaTCVAE [4.38301148531795]
This paper proposes a new iterative decomposition path of total correlation and explains the disentangled representation ability of VAE.
The novel model enables VAE to adjust the parameter capacity to divide dependent and independent data features flexibly.
arXiv Detail & Related papers (2022-10-17T07:16:53Z) - Revisiting GANs by Best-Response Constraint: Perspective, Methodology,
and Application [49.66088514485446]
Best-Response Constraint (BRC) is a general learning framework to explicitly formulate the potential dependency of the generator on the discriminator.
We show that even with different motivations and formulations, a variety of existing GANs ALL can be uniformly improved by our flexible BRC methodology.
arXiv Detail & Related papers (2022-05-20T12:42:41Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - InteL-VAEs: Adding Inductive Biases to Variational Auto-Encoders via
Intermediary Latents [60.785317191131284]
We introduce a simple and effective method for learning VAEs with controllable biases by using an intermediary set of latent variables.
In particular, it allows us to impose desired properties like sparsity or clustering on learned representations.
We show that this, in turn, allows InteL-VAEs to learn both better generative models and representations.
arXiv Detail & Related papers (2021-06-25T16:34:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.