Inference-InfoGAN: Inference Independence via Embedding Orthogonal Basis
Expansion
- URL: http://arxiv.org/abs/2110.00788v1
- Date: Sat, 2 Oct 2021 11:54:23 GMT
- Title: Inference-InfoGAN: Inference Independence via Embedding Orthogonal Basis
Expansion
- Authors: Hongxiang Jiang, Jihao Yin, Xiaoyan Luo, Fuxiang Wang
- Abstract summary: Disentanglement learning aims to construct independent and interpretable latent variables in which generative models are a popular strategy.
We propose a novel GAN-based disentanglement framework via embedding Orthogonal Basis Expansion (OBE) into InfoGAN network.
Our Inference-InfoGAN achieves higher disentanglement score in terms of FactorVAE, Separated ferenceAttribute Predictability (SAP), Mutual Information Gap (MIG) and Variation Predictability (VP) metrics without model fine-tuning.
- Score: 2.198430261120653
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Disentanglement learning aims to construct independent and interpretable
latent variables in which generative models are a popular strategy. InfoGAN is
a classic method via maximizing Mutual Information (MI) to obtain interpretable
latent variables mapped to the target space. However, it did not emphasize
independent characteristic. To explicitly infer latent variables with
inter-independence, we propose a novel GAN-based disentanglement framework via
embedding Orthogonal Basis Expansion (OBE) into InfoGAN network
(Inference-InfoGAN) in an unsupervised way. Under the OBE module, one set of
orthogonal basis can be adaptively found to expand arbitrary data with
independence property. To ensure the target-wise interpretable representation,
we add a consistence constraint between the expansion coefficients and latent
variables on the base of MI maximization. Additionally, we design an
alternating optimization step on the consistence constraint and orthogonal
requirement updating, so that the training of Inference-InfoGAN can be more
convenient. Finally, experiments validate that our proposed OBE module obtains
adaptive orthogonal basis, which can express better independent characteristics
than fixed basis expression of Discrete Cosine Transform (DCT). To depict the
performance in downstream tasks, we compared with the state-of-the-art
GAN-based and even VAE-based approaches on different datasets. Our
Inference-InfoGAN achieves higher disentanglement score in terms of FactorVAE,
Separated Attribute Predictability (SAP), Mutual Information Gap (MIG) and
Variation Predictability (VP) metrics without model fine-tuning. All the
experimental results illustrate that our method has inter-independence
inference ability because of the OBE module, and provides a good trade-off
between it and target-wise interpretability of latent variables via jointing
the alternating optimization.
Related papers
- Model-free Methods for Event History Analysis and Efficient Adjustment (PhD Thesis) [55.2480439325792]
This thesis is a series of independent contributions to statistics unified by a model-free perspective.
The first chapter elaborates on how a model-free perspective can be used to formulate flexible methods that leverage prediction techniques from machine learning.
The second chapter studies the concept of local independence, which describes whether the evolution of one process is directly influenced by another.
arXiv Detail & Related papers (2025-02-11T19:24:09Z) - Structural Entropy Guided Probabilistic Coding [52.01765333755793]
We propose a novel structural entropy-guided probabilistic coding model, named SEPC.
We incorporate the relationship between latent variables into the optimization by proposing a structural entropy regularization loss.
Experimental results across 12 natural language understanding tasks, including both classification and regression tasks, demonstrate the superior performance of SEPC.
arXiv Detail & Related papers (2024-12-12T00:37:53Z) - Uniform Transformation: Refining Latent Representation in Variational Autoencoders [7.4316292428754105]
We introduce a novel adaptable three-stage Uniform Transformation (UT) module to address irregular latent distributions.
By reconfiguring irregular distributions into a uniform distribution in the latent space, our approach significantly enhances the disentanglement and interpretability of latent representations.
Empirical evaluations demonstrated the efficacy of our proposed UT module in improving disentanglement metrics across benchmark datasets.
arXiv Detail & Related papers (2024-07-02T21:46:23Z) - Elastic Interaction Energy-Based Generative Model: Approximation in
Feature Space [14.783344918500813]
We propose a novel approach to generative modeling using a loss function based on elastic interaction energy (EIE)
The utilization of the EIE-based metric presents several advantages, including its long range property that enables consideration of global information in the distribution.
Experimental results on popular datasets, such as MNIST, FashionMNIST, CIFAR-10, and CelebA, demonstrate that our EIEG GAN model can mitigate mode collapse, enhance stability, and improve model performance.
arXiv Detail & Related papers (2023-03-19T03:39:31Z) - Variable Importance Matching for Causal Inference [73.25504313552516]
We describe a general framework called Model-to-Match that achieves these goals.
Model-to-Match uses variable importance measurements to construct a distance metric.
We operationalize the Model-to-Match framework with LASSO.
arXiv Detail & Related papers (2023-02-23T00:43:03Z) - Attitudes and Latent Class Choice Models using Machine learning [0.0]
We present a method of efficiently incorporating attitudinal indicators in the specification of Latent Class Choice Models (LCCM)
This formulation overcomes structural equations in its capability of exploring the relationship between the attitudinal indicators and the decision choice.
We test our proposed framework for estimating a Car-Sharing (CS) service subscription choice with stated preference data from Copenhagen, Denmark.
arXiv Detail & Related papers (2023-02-20T10:03:01Z) - Break The Spell Of Total Correlation In betaTCVAE [4.38301148531795]
This paper proposes a new iterative decomposition path of total correlation and explains the disentangled representation ability of VAE.
The novel model enables VAE to adjust the parameter capacity to divide dependent and independent data features flexibly.
arXiv Detail & Related papers (2022-10-17T07:16:53Z) - Revisiting GANs by Best-Response Constraint: Perspective, Methodology,
and Application [49.66088514485446]
Best-Response Constraint (BRC) is a general learning framework to explicitly formulate the potential dependency of the generator on the discriminator.
We show that even with different motivations and formulations, a variety of existing GANs ALL can be uniformly improved by our flexible BRC methodology.
arXiv Detail & Related papers (2022-05-20T12:42:41Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - InteL-VAEs: Adding Inductive Biases to Variational Auto-Encoders via
Intermediary Latents [60.785317191131284]
We introduce a simple and effective method for learning VAEs with controllable biases by using an intermediary set of latent variables.
In particular, it allows us to impose desired properties like sparsity or clustering on learned representations.
We show that this, in turn, allows InteL-VAEs to learn both better generative models and representations.
arXiv Detail & Related papers (2021-06-25T16:34:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.