Weakly Supervised Disentangled Generative Causal Representation Learning
- URL: http://arxiv.org/abs/2010.02637v3
- Date: Wed, 24 Aug 2022 12:57:57 GMT
- Title: Weakly Supervised Disentangled Generative Causal Representation Learning
- Authors: Xinwei Shen, Furui Liu, Hanze Dong, Qing Lian, Zhitang Chen, and Tong
Zhang
- Abstract summary: We show that previous methods with independent priors fail to disentangle causally related factors even under supervision.
We propose a new disentangled learning method that enables causal controllable generation and causal representation learning.
- Score: 21.392372783459013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a Disentangled gEnerative cAusal Representation (DEAR)
learning method under appropriate supervised information. Unlike existing
disentanglement methods that enforce independence of the latent variables, we
consider the general case where the underlying factors of interests can be
causally related. We show that previous methods with independent priors fail to
disentangle causally related factors even under supervision. Motivated by this
finding, we propose a new disentangled learning method called DEAR that enables
causal controllable generation and causal representation learning. The key
ingredient of this new formulation is to use a structural causal model (SCM) as
the prior distribution for a bidirectional generative model. The prior is then
trained jointly with a generator and an encoder using a suitable GAN algorithm
incorporated with supervised information on the ground-truth factors and their
underlying causal structure. We provide theoretical justification on the
identifiability and asymptotic convergence of the proposed method. We conduct
extensive experiments on both synthesized and real data sets to demonstrate the
effectiveness of DEAR in causal controllable generation, and the benefits of
the learned representations for downstream tasks in terms of sample efficiency
and distributional robustness.
Related papers
- Identifiable Latent Neural Causal Models [82.14087963690561]
Causal representation learning seeks to uncover latent, high-level causal representations from low-level observed data.
We determine the types of distribution shifts that do contribute to the identifiability of causal representations.
We translate our findings into a practical algorithm, allowing for the acquisition of reliable latent causal representations.
arXiv Detail & Related papers (2024-03-23T04:13:55Z) - Inducing Causal Structure for Abstractive Text Summarization [76.1000380429553]
We introduce a Structural Causal Model (SCM) to induce the underlying causal structure of the summarization data.
We propose a Causality Inspired Sequence-to-Sequence model (CI-Seq2Seq) to learn the causal representations that can mimic the causal factors.
Experimental results on two widely used text summarization datasets demonstrate the advantages of our approach.
arXiv Detail & Related papers (2023-08-24T16:06:36Z) - A Causal Ordering Prior for Unsupervised Representation Learning [27.18951912984905]
Causal representation learning argues that factors of variation in a dataset are, in fact, causally related.
We propose a fully unsupervised representation learning method that considers a data generation process with a latent additive noise model.
arXiv Detail & Related papers (2023-07-11T18:12:05Z) - Nonparametric Identifiability of Causal Representations from Unknown
Interventions [63.1354734978244]
We study causal representation learning, the task of inferring latent causal variables and their causal relations from mixtures of the variables.
Our goal is to identify both the ground truth latents and their causal graph up to a set of ambiguities which we show to be irresolvable from interventional data.
arXiv Detail & Related papers (2023-06-01T10:51:58Z) - Do-Operation Guided Causal Representation Learning with Reduced
Supervision Strength [12.012459418829732]
Causal representation learning has been proposed to encode relationships between factors presented in the high dimensional data.
We propose a framework which implements do-operation by swapping latent cause and effect factors encoded from a pair of inputs.
We also identify the inadequacy of existing causal representation metrics empirically and theoretically, and introduce new metrics for better evaluation.
arXiv Detail & Related papers (2022-06-03T20:18:04Z) - Principled Knowledge Extrapolation with GANs [92.62635018136476]
We study counterfactual synthesis from a new perspective of knowledge extrapolation.
We show that an adversarial game with a closed-form discriminator can be used to address the knowledge extrapolation problem.
Our method enjoys both elegant theoretical guarantees and superior performance in many scenarios.
arXiv Detail & Related papers (2022-05-21T08:39:42Z) - Causal Discovery in Linear Structural Causal Models with Deterministic
Relations [27.06618125828978]
We focus on the task of causal discovery form observational data.
We derive a set of necessary and sufficient conditions for unique identifiability of the causal structure.
arXiv Detail & Related papers (2021-10-30T21:32:42Z) - Learning Causal Semantic Representation for Out-of-Distribution
Prediction [125.38836464226092]
We propose a Causal Semantic Generative model (CSG) based on a causal reasoning so that the two factors are modeled separately.
We show that CSG can identify the semantic factor by fitting training data, and this semantic-identification guarantees the boundedness of OOD generalization error.
arXiv Detail & Related papers (2020-11-03T13:16:05Z) - CausalVAE: Structured Causal Disentanglement in Variational Autoencoder [52.139696854386976]
The framework of variational autoencoder (VAE) is commonly used to disentangle independent factors from observations.
We propose a new VAE based framework named CausalVAE, which includes a Causal Layer to transform independent factors into causal endogenous ones.
Results show that the causal representations learned by CausalVAE are semantically interpretable, and their causal relationship as a Directed Acyclic Graph (DAG) is identified with good accuracy.
arXiv Detail & Related papers (2020-04-18T20:09:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.