Differentiable Causal Discovery Under Unmeasured Confounding
- URL: http://arxiv.org/abs/2010.06978v2
- Date: Thu, 25 Feb 2021 02:20:26 GMT
- Title: Differentiable Causal Discovery Under Unmeasured Confounding
- Authors: Rohit Bhattacharya, Tushar Nagarajan, Daniel Malinsky, Ilya Shpitser
- Abstract summary: confounded systems exhibit more general equality restrictions that cannot be represented via ancestral ADMGs.
We use these constraints to cast causal discovery as a continuous optimization problem.
We demonstrate the efficacy of our method through simulations and application to a protein expression dataset.
- Score: 19.635669040319872
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The data drawn from biological, economic, and social systems are often
confounded due to the presence of unmeasured variables. Prior work in causal
discovery has focused on discrete search procedures for selecting acyclic
directed mixed graphs (ADMGs), specifically ancestral ADMGs, that encode
ordinary conditional independence constraints among the observed variables of
the system. However, confounded systems also exhibit more general equality
restrictions that cannot be represented via these graphs, placing a limit on
the kinds of structures that can be learned using ancestral ADMGs. In this
work, we derive differentiable algebraic constraints that fully characterize
the space of ancestral ADMGs, as well as more general classes of ADMGs, arid
ADMGs and bow-free ADMGs, that capture all equality restrictions on the
observed variables. We use these constraints to cast causal discovery as a
continuous optimization problem and design differentiable procedures to find
the best fitting ADMG when the data comes from a confounded linear system of
equations with correlated errors. We demonstrate the efficacy of our method
through simulations and application to a protein expression dataset. Code
implementing our methods is open-source and publicly available at
https://gitlab.com/rbhatta8/dcd and will be incorporated into the Ananke
package.
Related papers
- Analytic DAG Constraints for Differentiable DAG Learning [83.93320658222717]
We develop a theory to establish a connection between analytic functions and DAG constraints.
We show that analytic functions from the set $f(x) = c_0 + sum_i=1inftyc_ixi | forall i > 0, c_i > 0; r = lim_irightarrow inftyc_i/c_i+1 > 0$ can be employed to formulate effective DAG constraints.
arXiv Detail & Related papers (2025-03-24T23:51:35Z) - AcceleratedLiNGAM: Learning Causal DAGs at the speed of GPUs [57.12929098407975]
We show that by efficiently parallelizing existing causal discovery methods, we can scale them to thousands of dimensions.
Specifically, we focus on the causal ordering subprocedure in DirectLiNGAM and implement GPU kernels to accelerate it.
This allows us to apply DirectLiNGAM to causal inference on large-scale gene expression data with genetic interventions yielding competitive results.
arXiv Detail & Related papers (2024-03-06T15:06:11Z) - Toward the Identifiability of Comparative Deep Generative Models [7.5479347719819865]
We propose a theory of identifiability for comparative Deep Generative Models (DGMs)
We show that, while these models lack identifiability across a general class of mixing functions, they surprisingly become identifiable when the mixing function is piece-wise affine.
We also investigate the impact of model misspecification, and empirically show that previously proposed regularization techniques for fitting comparative DGMs help with identifiability when the number of latent variables is not known in advance.
arXiv Detail & Related papers (2024-01-29T06:10:54Z) - Dagma-DCE: Interpretable, Non-Parametric Differentiable Causal Discovery [19.939720617730472]
We introduce Dagma-DCE, an interpretable and model-agnostic scheme for differentiable causal discovery.
textscDagma-DCE uses an interpretable measure of causal strength to define weighted adjacency matrices.
arXiv Detail & Related papers (2024-01-05T18:15:19Z) - iSCAN: Identifying Causal Mechanism Shifts among Nonlinear Additive
Noise Models [48.33685559041322]
This paper focuses on identifying the causal mechanism shifts in two or more related datasets over the same set of variables.
Code implementing the proposed method is open-source and publicly available at https://github.com/kevinsbello/iSCAN.
arXiv Detail & Related papers (2023-06-30T01:48:11Z) - Modeling the Data-Generating Process is Necessary for Out-of-Distribution Generalization [23.302060306322506]
Real-world data often has multiple distribution shifts over different attributes.
No state-of-the-art DG algorithm performs consistently well on all shifts.
We develop Causally Adaptive Constraint Minimization (CACM), an algorithm that uses knowledge about the data-generating process to adaptively identify and apply the correct independence constraints for regularization.
arXiv Detail & Related papers (2022-06-15T22:35:06Z) - MissDAG: Causal Discovery in the Presence of Missing Data with
Continuous Additive Noise Models [78.72682320019737]
We develop a general method, which we call MissDAG, to perform causal discovery from data with incomplete observations.
MissDAG maximizes the expected likelihood of the visible part of observations under the expectation-maximization framework.
We demonstrate the flexibility of MissDAG for incorporating various causal discovery algorithms and its efficacy through extensive simulations and real data experiments.
arXiv Detail & Related papers (2022-05-27T09:59:46Z) - Data-Efficient and Interpretable Tabular Anomaly Detection [54.15249463477813]
We propose a novel framework that adapts a white-box model class, Generalized Additive Models, to detect anomalies.
In addition, the proposed framework, DIAD, can incorporate a small amount of labeled data to further boost anomaly detection performances in semi-supervised settings.
arXiv Detail & Related papers (2022-03-03T22:02:56Z) - BCDAG: An R package for Bayesian structure and Causal learning of
Gaussian DAGs [77.34726150561087]
We introduce the R package for causal discovery and causal effect estimation from observational data.
Our implementation scales efficiently with the number of observations and, whenever the DAGs are sufficiently sparse, the number of variables in the dataset.
We then illustrate the main functions and algorithms on both real and simulated datasets.
arXiv Detail & Related papers (2022-01-28T09:30:32Z) - Causal Discovery in Knowledge Graphs by Exploiting Asymmetric Properties
of Non-Gaussian Distributions [3.1981440103815717]
We define a hybrid approach that allows us to discover cause-effect relationships in Knowledge Graphs.
The proposed approach is based around the finding of the instantaneous causal structure of a non-experimental matrix using a non-Gaussian model.
We use two different pre-existing algorithms, one for the causal discovery and the other for decomposing the Knowledge Graph.
arXiv Detail & Related papers (2021-06-02T09:33:05Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.