DeCaFlow: A Deconfounding Causal Generative Model
- URL: http://arxiv.org/abs/2503.15114v2
- Date: Sat, 24 May 2025 12:18:46 GMT
- Title: DeCaFlow: A Deconfounding Causal Generative Model
- Authors: Alejandro Almodóvar, Adrián Javaloy, Juan Parras, Santiago Zazo, Isabel Valera,
- Abstract summary: We introduce DeCaFlow, a deconfounding causal generative model.<n>We extend previous results on causal estimation under hidden confounding to show that a single instance of DeCaFlow provides correct estimates for all causal queries identifiable with do-calculus.<n>Our empirical results on diverse settings show that DeCaFlow outperforms existing approaches, while demonstrating its out-of-the-box applicability to any given causal graph.
- Score: 58.411886466157185
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We introduce DeCaFlow, a deconfounding causal generative model. Training once per dataset using just observational data and the underlying causal graph, DeCaFlow enables accurate causal inference on continuous variables under the presence of hidden confounders. Specifically, we extend previous results on causal estimation under hidden confounding to show that a single instance of DeCaFlow provides correct estimates for all causal queries identifiable with do-calculus, leveraging proxy variables to adjust for the causal effects when do-calculus alone is insufficient. Moreover, we show that counterfactual queries are identifiable as long as their interventional counterparts are identifiable, and thus are also correctly estimated by DeCaFlow. Our empirical results on diverse settings (including the Ecoli70 dataset, with 3 independent hidden confounders, tens of observed variables and hundreds of causal queries) show that DeCaFlow outperforms existing approaches, while demonstrating its out-of-the-box applicability to any given causal graph. An implementation can be found in https://github.com/aalmodovares/DeCaFlow
Related papers
- Clustering and Pruning in Causal Data Fusion [1.0923877073891441]
Do-calculus remains the only general-purpose tool for causal data fusion.<n>We propose pruning (removing unnecessary variables) and clustering (combining variables) as preprocessing operations.
arXiv Detail & Related papers (2025-05-21T07:44:39Z) - Estimating Causal Effects from Learned Causal Networks [56.14597641617531]
We propose an alternative paradigm for answering causal-effect queries over discrete observable variables.
We learn the causal Bayesian network and its confounding latent variables directly from the observational data.
We show that this emphmodel completion learning approach can be more effective than estimand approaches.
arXiv Detail & Related papers (2024-08-26T08:39:09Z) - Detection of Unobserved Common Causes based on NML Code in Discrete,
Mixed, and Continuous Variables [1.5039745292757667]
We categorize all possible causal relationships between two random variables into the following four categories.
We show that CLOUD is more effective than existing methods in inferring causal relationships by extensive experiments on both synthetic and real-world data.
arXiv Detail & Related papers (2024-03-11T08:11:52Z) - Federated Causal Discovery from Heterogeneous Data [70.31070224690399]
We propose a novel FCD method attempting to accommodate arbitrary causal models and heterogeneous data.
These approaches involve constructing summary statistics as a proxy of the raw data to protect data privacy.
We conduct extensive experiments on synthetic and real datasets to show the efficacy of our method.
arXiv Detail & Related papers (2024-02-20T18:53:53Z) - Sample, estimate, aggregate: A recipe for causal discovery foundation models [28.116832159265964]
Causal discovery has the potential to uncover mechanistic insights from biological experiments.
We propose a supervised model trained on large-scale, synthetic data to predict causal graphs.
Our approach is enabled by the observation that typical errors in the outputs of a discovery algorithm remain comparable across datasets.
arXiv Detail & Related papers (2024-02-02T21:57:58Z) - Causal Representation Learning Made Identifiable by Grouping of Observational Variables [8.157856010838382]
Causal Representation Learning aims to learn a causal model for hidden features in a data-driven manner.
Here, we show identifiability based on novel, weak constraints.
We also propose a novel self-supervised estimation framework consistent with the model.
arXiv Detail & Related papers (2023-10-24T10:38:02Z) - Causal normalizing flows: from theory to practice [10.733905678329675]
We use recent results on non-linear ICA to show that causal models are identifiable from observational data given a causal ordering.
Second, we analyze different design and learning choices for causal normalizing flows to capture the underlying causal data-generating process.
Third, we describe how to implement the do-operator in causal NFs, and thus, how to answer interventional and counterfactual questions.
arXiv Detail & Related papers (2023-06-08T17:58:05Z) - Causal Component Analysis [20.570470860880125]
We introduce an intermediate problem termed Causal Component Analysis (CauCA)
CauCA can be viewed as a generalization of Independent Component Analysis (ICA)
We demonstrate its effectiveness through extensive synthetic experiments in the CauCA and ICA setting.
arXiv Detail & Related papers (2023-05-26T19:34:35Z) - Sequential Causal Effect Variational Autoencoder: Time Series Causal
Link Estimation under Hidden Confounding [8.330791157878137]
Estimating causal effects from observational data sometimes leads to spurious relationships which can be misconceived as causal.
We propose Sequential Causal Effect Variational Autoencoder (SCEVAE), a novel method for time series causality analysis under hidden confounding.
arXiv Detail & Related papers (2022-09-23T09:43:58Z) - Active Bayesian Causal Inference [72.70593653185078]
We propose Active Bayesian Causal Inference (ABCI), a fully-Bayesian active learning framework for integrated causal discovery and reasoning.
ABCI jointly infers a posterior over causal models and queries of interest.
We show that our approach is more data-efficient than several baselines that only focus on learning the full causal graph.
arXiv Detail & Related papers (2022-06-04T22:38:57Z) - Effect Identification in Cluster Causal Diagrams [51.42809552422494]
We introduce a new type of graphical model called cluster causal diagrams (for short, C-DAGs)
C-DAGs allow for the partial specification of relationships among variables based on limited prior knowledge.
We develop the foundations and machinery for valid causal inferences over C-DAGs.
arXiv Detail & Related papers (2022-02-22T21:27:31Z) - Disentangling Observed Causal Effects from Latent Confounders using
Method of Moments [67.27068846108047]
We provide guarantees on identifiability and learnability under mild assumptions.
We develop efficient algorithms based on coupled tensor decomposition with linear constraints to obtain scalable and guaranteed solutions.
arXiv Detail & Related papers (2021-01-17T07:48:45Z) - Causal Expectation-Maximisation [70.45873402967297]
We show that causal inference is NP-hard even in models characterised by polytree-shaped graphs.
We introduce the causal EM algorithm to reconstruct the uncertainty about the latent variables from data about categorical manifest variables.
We argue that there appears to be an unnoticed limitation to the trending idea that counterfactual bounds can often be computed without knowledge of the structural equations.
arXiv Detail & Related papers (2020-11-04T10:25:13Z) - Structural Causal Models Are (Solvable by) Credal Networks [70.45873402967297]
Causal inferences can be obtained by standard algorithms for the updating of credal nets.
This contribution should be regarded as a systematic approach to represent structural causal models by credal networks.
Experiments show that approximate algorithms for credal networks can immediately be used to do causal inference in real-size problems.
arXiv Detail & Related papers (2020-08-02T11:19:36Z) - Amortized Causal Discovery: Learning to Infer Causal Graphs from
Time-Series Data [63.15776078733762]
We propose Amortized Causal Discovery, a novel framework to learn to infer causal relations from time-series data.
We demonstrate experimentally that this approach, implemented as a variational model, leads to significant improvements in causal discovery performance.
arXiv Detail & Related papers (2020-06-18T19:59:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.