DoWhy: Addressing Challenges in Expressing and Validating Causal
Assumptions
- URL: http://arxiv.org/abs/2108.13518v1
- Date: Fri, 27 Aug 2021 11:07:30 GMT
- Title: DoWhy: Addressing Challenges in Expressing and Validating Causal
Assumptions
- Authors: Amit Sharma, Vasilis Syrgkanis, Cheng Zhang, Emre K{\i}c{\i}man
- Abstract summary: DoWhy is a framework that allows explicit declaration of assumptions through a causal graph.
It provides multiple validation tests to check a subset of these assumptions.
Our experience with DoWhy highlights a number of open questions for future research.
- Score: 40.70930937915354
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Estimation of causal effects involves crucial assumptions about the
data-generating process, such as directionality of effect, presence of
instrumental variables or mediators, and whether all relevant confounders are
observed. Violation of any of these assumptions leads to significant error in
the effect estimate. However, unlike cross-validation for predictive models,
there is no global validator method for a causal estimate. As a result,
expressing different causal assumptions formally and validating them (to the
extent possible) becomes critical for any analysis. We present DoWhy, a
framework that allows explicit declaration of assumptions through a causal
graph and provides multiple validation tests to check a subset of these
assumptions. Our experience with DoWhy highlights a number of open questions
for future research: developing new ways beyond causal graphs to express
assumptions, the role of causal discovery in learning relevant parts of the
graph, and developing validation tests that can better detect errors, both for
average and conditional treatment effects. DoWhy is available at
https://github.com/microsoft/dowhy.
Related papers
- Data Fusion for Partial Identification of Causal Effects [62.56890808004615]
We propose a novel partial identification framework that enables researchers to answer key questions.<n>Is the causal effect positive or negative? and How severe must assumption violations be to overturn this conclusion?<n>We apply our framework to the Project STAR study, which investigates the effect of classroom size on students' third-grade standardized test performance.
arXiv Detail & Related papers (2025-05-30T07:13:01Z) - Causal Inference Isn't Special: Why It's Just Another Prediction Problem [1.90365714903665]
Causal inference is often portrayed as distinct from predictive modeling.
But at its core, causal inference is simply a structured instance of prediction under distribution shift.
This perspective reframes causal estimation as a familiar generalization problem.
arXiv Detail & Related papers (2025-04-06T01:37:50Z) - Score matching through the roof: linear, nonlinear, and latent variables causal discovery [18.46845413928147]
Causal discovery from observational data holds great promise.
Existing methods rely on strong assumptions about the underlying causal structure.
We propose a flexible algorithm for causal discovery across linear, nonlinear, and latent variable models.
arXiv Detail & Related papers (2024-07-26T14:09:06Z) - Self-Compatibility: Evaluating Causal Discovery without Ground Truth [28.72650348646176]
We propose a novel method for falsifying the output of a causal discovery algorithm in the absence of ground truth.
Our key insight is that while statistical learning seeks stability across subsets of data points, causal learning should seek stability across subsets of variables.
We prove that detecting incompatibilities can falsify wrongly inferred causal relations due to violation of assumptions or errors from finite sample effects.
arXiv Detail & Related papers (2023-07-18T18:59:42Z) - Nonparametric Identifiability of Causal Representations from Unknown
Interventions [63.1354734978244]
We study causal representation learning, the task of inferring latent causal variables and their causal relations from mixtures of the variables.
Our goal is to identify both the ground truth latents and their causal graph up to a set of ambiguities which we show to be irresolvable from interventional data.
arXiv Detail & Related papers (2023-06-01T10:51:58Z) - Active Bayesian Causal Inference [72.70593653185078]
We propose Active Bayesian Causal Inference (ABCI), a fully-Bayesian active learning framework for integrated causal discovery and reasoning.
ABCI jointly infers a posterior over causal models and queries of interest.
We show that our approach is more data-efficient than several baselines that only focus on learning the full causal graph.
arXiv Detail & Related papers (2022-06-04T22:38:57Z) - Causal Transportability for Visual Recognition [70.13627281087325]
We show that standard classifiers fail because the association between images and labels is not transportable across settings.
We then show that the causal effect, which severs all sources of confounding, remains invariant across domains.
This motivates us to develop an algorithm to estimate the causal effect for image classification.
arXiv Detail & Related papers (2022-04-26T15:02:11Z) - Typing assumptions improve identification in causal discovery [123.06886784834471]
Causal discovery from observational data is a challenging task to which an exact solution cannot always be identified.
We propose a new set of assumptions that constrain possible causal relationships based on the nature of the variables.
arXiv Detail & Related papers (2021-07-22T14:23:08Z) - DoWhy: An End-to-End Library for Causal Inference [16.764873959182765]
We describe DoWhy, an open-source Python library that is built with causal assumptions as its first-class citizens.
DoWhy presents an API for the four steps common to any causal analysis.
In particular, DoWhy implements a number of checks including placebo tests, bootstrap tests, and tests for unoberved confounding.
arXiv Detail & Related papers (2020-11-09T06:22:11Z) - Causal Expectation-Maximisation [70.45873402967297]
We show that causal inference is NP-hard even in models characterised by polytree-shaped graphs.
We introduce the causal EM algorithm to reconstruct the uncertainty about the latent variables from data about categorical manifest variables.
We argue that there appears to be an unnoticed limitation to the trending idea that counterfactual bounds can often be computed without knowledge of the structural equations.
arXiv Detail & Related papers (2020-11-04T10:25:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.