Evaluation of Induced Expert Knowledge in Causal Structure Learning by
NOTEARS
- URL: http://arxiv.org/abs/2301.01817v1
- Date: Wed, 4 Jan 2023 20:39:39 GMT
- Title: Evaluation of Induced Expert Knowledge in Causal Structure Learning by
NOTEARS
- Authors: Jawad Chowdhury, Rezaur Rashid, Gabriel Terejanu
- Abstract summary: We study the impact of expert knowledge on causal relations in the form of additional constraints used in the formulation of the nonparametric NOTEARS model.
We found that (i) knowledge that corrects the mistakes of the NOTEARS model can lead to statistically significant improvements, (ii) constraints on active edges have a larger positive impact on causal discovery than inactive edges, and surprisingly, (iii) the induced knowledge does not correct on average more incorrect active and/or inactive edges than expected.
- Score: 1.5469452301122175
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Causal modeling provides us with powerful counterfactual reasoning and
interventional mechanism to generate predictions and reason under various
what-if scenarios. However, causal discovery using observation data remains a
nontrivial task due to unobserved confounding factors, finite sampling, and
changes in the data distribution. These can lead to spurious cause-effect
relationships. To mitigate these challenges in practice, researchers augment
causal learning with known causal relations. The goal of the paper is to study
the impact of expert knowledge on causal relations in the form of additional
constraints used in the formulation of the nonparametric NOTEARS. We provide a
comprehensive set of comparative analyses of biasing the model using different
types of knowledge. We found that (i) knowledge that corrects the mistakes of
the NOTEARS model can lead to statistically significant improvements, (ii)
constraints on active edges have a larger positive impact on causal discovery
than inactive edges, and surprisingly, (iii) the induced knowledge does not
correct on average more incorrect active and/or inactive edges than expected.
We also demonstrate the behavior of the model and the effectiveness of domain
knowledge on a real-world dataset.
Related papers
- DAG-aware Transformer for Causal Effect Estimation [0.8192907805418583]
Causal inference is a critical task across fields such as healthcare, economics, and the social sciences.
In this paper, we present a novel transformer-based method for causal inference that overcomes these challenges.
The core innovation of our model lies in its integration of causal Directed Acyclic Graphs (DAGs) directly into the attention mechanism.
arXiv Detail & Related papers (2024-10-13T23:17:58Z) - From Pre-training Corpora to Large Language Models: What Factors Influence LLM Performance in Causal Discovery Tasks? [51.42906577386907]
This study explores the factors influencing the performance of Large Language Models (LLMs) in causal discovery tasks.
A higher frequency of causal mentions correlates with better model performance, suggesting that extensive exposure to causal information during training enhances the models' causal discovery capabilities.
arXiv Detail & Related papers (2024-07-29T01:45:05Z) - Identifiable Latent Polynomial Causal Models Through the Lens of Change [82.14087963690561]
Causal representation learning aims to unveil latent high-level causal representations from observed low-level data.
One of its primary tasks is to provide reliable assurance of identifying these latent causal models, known as identifiability.
arXiv Detail & Related papers (2023-10-24T07:46:10Z) - Interpretable Imitation Learning with Dynamic Causal Relations [65.18456572421702]
We propose to expose captured knowledge in the form of a directed acyclic causal graph.
We also design this causal discovery process to be state-dependent, enabling it to model the dynamics in latent causal graphs.
The proposed framework is composed of three parts: a dynamic causal discovery module, a causality encoding module, and a prediction module, and is trained in an end-to-end manner.
arXiv Detail & Related papers (2023-09-30T20:59:42Z) - Advancing Counterfactual Inference through Nonlinear Quantile Regression [77.28323341329461]
We propose a framework for efficient and effective counterfactual inference implemented with neural networks.
The proposed approach enhances the capacity to generalize estimated counterfactual outcomes to unseen data.
Empirical results conducted on multiple datasets offer compelling support for our theoretical assertions.
arXiv Detail & Related papers (2023-06-09T08:30:51Z) - Quantify the Causes of Causal Emergence: Critical Conditions of
Uncertainty and Asymmetry in Causal Structure [0.5372002358734439]
Investigation of causal relationships based on statistical and informational theories have posed an interesting and valuable challenge to large-scale models.
This paper introduces a framework for assessing numerical conditions of Causal Emergence as theoretical constraints of its occurrence.
arXiv Detail & Related papers (2022-12-03T06:35:54Z) - Active Bayesian Causal Inference [72.70593653185078]
We propose Active Bayesian Causal Inference (ABCI), a fully-Bayesian active learning framework for integrated causal discovery and reasoning.
ABCI jointly infers a posterior over causal models and queries of interest.
We show that our approach is more data-efficient than several baselines that only focus on learning the full causal graph.
arXiv Detail & Related papers (2022-06-04T22:38:57Z) - Causal Knowledge Guided Societal Event Forecasting [24.437437565689393]
We introduce a deep learning framework that integrates causal effect estimation into event forecasting.
Two robust learning modules, including a feature reweighting module and an approximate loss, are introduced to enable prior knowledge injection.
arXiv Detail & Related papers (2021-12-10T17:41:02Z) - A Critical Look At The Identifiability of Causal Effects with Deep
Latent Variable Models [2.326384409283334]
We use causal effect variational autoencoder (CEVAE) as a case study.
CEVAE seems to work reliably under some simple scenarios, but it does not identify the correct causal effect with a misspecified latent variable or a complex data distribution.
Our results show that the question of identifiability cannot be disregarded, and we argue that more attention should be paid to it in future work.
arXiv Detail & Related papers (2021-02-12T17:43:18Z) - Influence Functions in Deep Learning Are Fragile [52.31375893260445]
influence functions approximate the effect of samples in test-time predictions.
influence estimates are fairly accurate for shallow networks.
Hessian regularization is important to get highquality influence estimates.
arXiv Detail & Related papers (2020-06-25T18:25:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.