Causal Inference with Conditional Front-Door Adjustment and Identifiable
Variational Autoencoder
- URL: http://arxiv.org/abs/2310.01937v1
- Date: Tue, 3 Oct 2023 10:24:44 GMT
- Title: Causal Inference with Conditional Front-Door Adjustment and Identifiable
Variational Autoencoder
- Authors: Ziqi Xu, Debo Cheng, Jiuyong Li, Jixue Liu, Lin Liu, Kui Yu
- Abstract summary: Front-door adjustment is a practical approach for dealing with unobserved confounding variables.
We develop the theorem that guarantees the causal effect identifiability of CFD adjustment.
We propose CFDiVAE to learn the representation of the CFD adjustment variable directly from data.
- Score: 28.94606676886985
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An essential and challenging problem in causal inference is causal effect
estimation from observational data. The problem becomes more difficult with the
presence of unobserved confounding variables. The front-door adjustment is a
practical approach for dealing with unobserved confounding variables. However,
the restriction for the standard front-door adjustment is difficult to satisfy
in practice. In this paper, we relax some of the restrictions by proposing the
concept of conditional front-door (CFD) adjustment and develop the theorem that
guarantees the causal effect identifiability of CFD adjustment. Furthermore, as
it is often impossible for a CFD variable to be given in practice, it is
desirable to learn it from data. By leveraging the ability of deep generative
models, we propose CFDiVAE to learn the representation of the CFD adjustment
variable directly from data with the identifiable Variational AutoEncoder and
formally prove the model identifiability. Extensive experiments on synthetic
datasets validate the effectiveness of CFDiVAE and its superiority over
existing methods. The experiments also show that the performance of CFDiVAE is
less sensitive to the causal strength of unobserved confounding variables. We
further apply CFDiVAE to a real-world dataset to demonstrate its potential
application.
Related papers
- Rectified Diffusion Guidance for Conditional Generation [62.00207951161297]
We revisit the theory behind CFG and rigorously confirm that the improper configuration of the combination coefficients (i.e., the widely used summing-to-one version) brings about expectation shift of the generative distribution.
We propose ReCFG with a relaxation on the guidance coefficients such that denoising with ReCFG strictly aligns with the diffusion theory.
That way the rectified coefficients can be readily pre-computed via traversing the observed data, leaving the sampling speed barely affected.
arXiv Detail & Related papers (2024-10-24T13:41:32Z) - Topic-aware Causal Intervention for Counterfactual Detection [7.295153420065765]
We consider the problem of counterfactual detection (CFD) and seek to enhance the CFD models.
Previous models are reliant on clue phrases to predict counterfactuality, so they suffer from significant performance drop when clue phrase hints do not exist during testing.
We propose to integrate neural topic model into the CFD model to capture the global semantics of the input statement.
arXiv Detail & Related papers (2024-09-25T06:55:33Z) - Causal Effect Estimation using identifiable Variational AutoEncoder with Latent Confounders and Post-Treatment Variables [18.34462010115951]
Estimating causal effects from observational data is challenging, especially in the presence of latent confounders.
We propose a novel method of joint Variational AutoEncoder (VAE) and identifiable Variational AutoEncoder (iVAE) for learning the representations of latent confounders and latent post-treatment variables.
arXiv Detail & Related papers (2024-08-13T22:13:25Z) - COD: Learning Conditional Invariant Representation for Domain Adaptation Regression [20.676363400841495]
Domain Adaptation Regression is developed to generalize label knowledge from a source domain to an unlabeled target domain.
Existing conditional distribution alignment theory and methods with discrete prior are no longer applicable.
To minimize the discrepancy, a COD-based conditional invariant representation learning model is proposed.
arXiv Detail & Related papers (2024-08-13T05:08:13Z) - Federated Causal Discovery from Heterogeneous Data [70.31070224690399]
We propose a novel FCD method attempting to accommodate arbitrary causal models and heterogeneous data.
These approaches involve constructing summary statistics as a proxy of the raw data to protect data privacy.
We conduct extensive experiments on synthetic and real datasets to show the efficacy of our method.
arXiv Detail & Related papers (2024-02-20T18:53:53Z) - Efficient Conformal Prediction under Data Heterogeneity [79.35418041861327]
Conformal Prediction (CP) stands out as a robust framework for uncertainty quantification.
Existing approaches for tackling non-exchangeability lead to methods that are not computable beyond the simplest examples.
This work introduces a new efficient approach to CP that produces provably valid confidence sets for fairly general non-exchangeable data distributions.
arXiv Detail & Related papers (2023-12-25T20:02:51Z) - Advancing Counterfactual Inference through Nonlinear Quantile Regression [77.28323341329461]
We propose a framework for efficient and effective counterfactual inference implemented with neural networks.
The proposed approach enhances the capacity to generalize estimated counterfactual outcomes to unseen data.
Empirical results conducted on multiple datasets offer compelling support for our theoretical assertions.
arXiv Detail & Related papers (2023-06-09T08:30:51Z) - Federated Conformal Predictors for Distributed Uncertainty
Quantification [83.50609351513886]
Conformal prediction is emerging as a popular paradigm for providing rigorous uncertainty quantification in machine learning.
In this paper, we extend conformal prediction to the federated learning setting.
We propose a weaker notion of partial exchangeability, better suited to the FL setting, and use it to develop the Federated Conformal Prediction framework.
arXiv Detail & Related papers (2023-05-27T19:57:27Z) - Causal Effect Estimation with Variational AutoEncoder and the Front Door
Criterion [23.20371860838245]
The front-door criterion is often difficult to identify the set of variables used for front-door adjustment from data.
By leveraging the ability of deep generative models in representation learning, we propose FDVAE to learn the representation of a Front-Door adjustment set with a Variational AutoEncoder.
arXiv Detail & Related papers (2023-04-24T10:04:28Z) - Diffusion Causal Models for Counterfactual Estimation [18.438307666925425]
We consider the task of counterfactual estimation from observational imaging data given a known causal structure.
We propose Diff-SCM, a deep structural causal model that builds on recent advances of generative energy-based models.
We find that Diff-SCM produces more realistic and minimal counterfactuals than baselines on MNIST data and can also be applied to ImageNet data.
arXiv Detail & Related papers (2022-02-21T12:23:01Z) - Variational Causal Networks: Approximate Bayesian Inference over Causal
Structures [132.74509389517203]
We introduce a parametric variational family modelled by an autoregressive distribution over the space of discrete DAGs.
In experiments, we demonstrate that the proposed variational posterior is able to provide a good approximation of the true posterior.
arXiv Detail & Related papers (2021-06-14T17:52:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.