Towards Causal Representation Learning and Deconfounding from Indefinite
Data
- URL: http://arxiv.org/abs/2305.02640v4
- Date: Fri, 11 Aug 2023 09:30:08 GMT
- Title: Towards Causal Representation Learning and Deconfounding from Indefinite
Data
- Authors: Hang Chen and Xinyu Yang and Qing Yang
- Abstract summary: Non-statistical data (e.g., images, text, etc.) encounters significant conflicts in terms of properties and methods with traditional causal data.
We redefine causal data from two novel perspectives and then propose three data paradigms.
We implement the above designs as a dynamic variational inference model, tailored to learn causal representation from indefinite data.
- Score: 17.793702165499298
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Owing to the cross-pollination between causal discovery and deep learning,
non-statistical data (e.g., images, text, etc.) encounters significant
conflicts in terms of properties and methods with traditional causal data. To
unify these data types of varying forms, we redefine causal data from two novel
perspectives and then propose three data paradigms. Among them, the indefinite
data (like dialogues or video sources) induce low sample utilization and
incapability of the distribution assumption, both leading to the fact that
learning causal representation from indefinite data is, as of yet, largely
unexplored. We design the causal strength variational model to settle down
these two problems. Specifically, we leverage the causal strength instead of
independent noise as the latent variable to construct evidence lower bound. By
this design ethos, The causal strengths of different structures are regarded as
a distribution and can be expressed as a 2D matrix. Moreover, considering the
latent confounders, we disentangle the causal graph G into two relation
subgraphs O and C. O contains pure relations between observed variables, while
C represents the relations from latent variables to observed variables. We
implement the above designs as a dynamic variational inference model, tailored
to learn causal representation from indefinite data under latent confounding.
Finally, we conduct comprehensive experiments on synthetic and real-world data
to demonstrate the effectiveness of our method.
Related papers
- Towards Causal Relationship in Indefinite Data: Baseline Model and New
Datasets [23.035761299444953]
"Indefinite Data" is characterized by multi-structure data and multi-value representations.
We release two high-quality datasets - Causalogue and Causaction.
We propose a probabilistic framework as a baseline, incorporating three designed highlights for this gap.
arXiv Detail & Related papers (2024-01-16T09:15:43Z) - Identifiable Latent Polynomial Causal Models Through the Lens of Change [82.14087963690561]
Causal representation learning aims to unveil latent high-level causal representations from observed low-level data.
One of its primary tasks is to provide reliable assurance of identifying these latent causal models, known as identifiability.
arXiv Detail & Related papers (2023-10-24T07:46:10Z) - A Causal Ordering Prior for Unsupervised Representation Learning [27.18951912984905]
Causal representation learning argues that factors of variation in a dataset are, in fact, causally related.
We propose a fully unsupervised representation learning method that considers a data generation process with a latent additive noise model.
arXiv Detail & Related papers (2023-07-11T18:12:05Z) - Nonparametric Identifiability of Causal Representations from Unknown
Interventions [63.1354734978244]
We study causal representation learning, the task of inferring latent causal variables and their causal relations from mixtures of the variables.
Our goal is to identify both the ground truth latents and their causal graph up to a set of ambiguities which we show to be irresolvable from interventional data.
arXiv Detail & Related papers (2023-06-01T10:51:58Z) - Identifying Weight-Variant Latent Causal Models [82.14087963690561]
We find that transitivity acts as a key role in impeding the identifiability of latent causal representations.
Under some mild assumptions, we can show that the latent causal representations can be identified up to trivial permutation and scaling.
We propose a novel method, termed Structural caUsAl Variational autoEncoder, which directly learns latent causal representations and causal relationships among them.
arXiv Detail & Related papers (2022-08-30T11:12:59Z) - Variational Causal Networks: Approximate Bayesian Inference over Causal
Structures [132.74509389517203]
We introduce a parametric variational family modelled by an autoregressive distribution over the space of discrete DAGs.
In experiments, we demonstrate that the proposed variational posterior is able to provide a good approximation of the true posterior.
arXiv Detail & Related papers (2021-06-14T17:52:49Z) - On Disentangled Representations Learned From Correlated Data [59.41587388303554]
We bridge the gap to real-world scenarios by analyzing the behavior of the most prominent disentanglement approaches on correlated data.
We show that systematically induced correlations in the dataset are being learned and reflected in the latent representations.
We also demonstrate how to resolve these latent correlations, either using weak supervision during training or by post-hoc correcting a pre-trained model with a small number of labels.
arXiv Detail & Related papers (2020-06-14T12:47:34Z) - CausalVAE: Structured Causal Disentanglement in Variational Autoencoder [52.139696854386976]
The framework of variational autoencoder (VAE) is commonly used to disentangle independent factors from observations.
We propose a new VAE based framework named CausalVAE, which includes a Causal Layer to transform independent factors into causal endogenous ones.
Results show that the causal representations learned by CausalVAE are semantically interpretable, and their causal relationship as a Directed Acyclic Graph (DAG) is identified with good accuracy.
arXiv Detail & Related papers (2020-04-18T20:09:34Z) - Causal discovery of linear non-Gaussian acyclic models in the presence
of latent confounders [6.1221613913018675]
This paper proposes a causal functional model-based method called repetitive causal discovery (RCD) to discover the causal structure of observed variables affected by latent confounders.
RCD repeats inferring the causal directions between a small number of observed variables and determines whether the relationships are affected by latent confounders.
arXiv Detail & Related papers (2020-01-13T12:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.