Deep Structural Causal Models for Tractable Counterfactual Inference
- URL: http://arxiv.org/abs/2006.06485v2
- Date: Thu, 22 Oct 2020 21:59:47 GMT
- Title: Deep Structural Causal Models for Tractable Counterfactual Inference
- Authors: Nick Pawlowski, Daniel C. Castro, Ben Glocker
- Abstract summary: We formulate a general framework for building structural causal models (SCMs) with deep learning components.
Our framework is validated on a synthetic dataset built on MNIST as well as on a real-world medical dataset of brain MRI scans.
- Score: 24.26709730032233
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We formulate a general framework for building structural causal models (SCMs)
with deep learning components. The proposed approach employs normalising flows
and variational inference to enable tractable inference of exogenous noise
variables - a crucial step for counterfactual inference that is missing from
existing deep causal learning methods. Our framework is validated on a
synthetic dataset built on MNIST as well as on a real-world medical dataset of
brain MRI scans. Our experimental results indicate that we can successfully
train deep SCMs that are capable of all three levels of Pearl's ladder of
causation: association, intervention, and counterfactuals, giving rise to a
powerful new approach for answering causal questions in imaging applications
and beyond. The code for all our experiments is available at
https://github.com/biomedia-mira/deepscm.
Related papers
- MindBridge: A Cross-Subject Brain Decoding Framework [60.58552697067837]
Brain decoding aims to reconstruct stimuli from acquired brain signals.
Currently, brain decoding is confined to a per-subject-per-model paradigm.
We present MindBridge, that achieves cross-subject brain decoding by employing only one model.
arXiv Detail & Related papers (2024-04-11T15:46:42Z) - Discovery of the Hidden World with Large Language Models [95.58823685009727]
This paper presents Causal representatiOn AssistanT (COAT) that introduces large language models (LLMs) to bridge the gap.
LLMs are trained on massive observations of the world and have demonstrated great capability in extracting key information from unstructured data.
COAT also adopts CDs to find causal relations among the identified variables as well as to provide feedback to LLMs to iteratively refine the proposed factors.
arXiv Detail & Related papers (2024-02-06T12:18:54Z) - Causal Optimal Transport of Abstractions [8.642152250082368]
Causal abstraction (CA) theory establishes formal criteria for relating multiple structural causal models (SCMs) at different levels of granularity.
We propose COTA, the first method to learn abstraction maps from observational and interventional data without assuming complete knowledge of the underlying SCMs.
We extensively evaluate COTA on synthetic and real world problems, and showcase its advantages over non-causal, independent and aggregated COTA formulations.
arXiv Detail & Related papers (2023-12-13T12:54:34Z) - How to Enhance Causal Discrimination of Utterances: A Case on Affective
Reasoning [22.11437627661179]
We propose the incorporation of textiti.i.i.d. noise terms into the conversation process, thereby constructing a structural causal model (SCM)
To facilitate the implementation of deep learning, we introduce the cogn frameworks to handle unstructured conversation data, and employ an autoencoder architecture to regard the unobservable noise as learnable "implicit causes"
arXiv Detail & Related papers (2023-05-04T07:45:49Z) - Learning Latent Structural Causal Models [31.686049664958457]
In machine learning tasks, one often operates on low-level data like image pixels or high-dimensional vectors.
We present a tractable approximate inference method which performs joint inference over the causal variables, structure and parameters of the latent Structural Causal Model.
arXiv Detail & Related papers (2022-10-24T20:09:44Z) - On the Robustness and Generalization of Deep Learning Driven Full
Waveform Inversion [2.5382095320488665]
Full Waveform Inversion (FWI) is commonly epitomized as an image-to-image translation task.
Despite being trained with synthetic data, the deep learning-driven FWI is expected to perform well when evaluated with sufficient real-world data.
We study such properties by asking: how robust are these deep neural networks and how do they generalize?
arXiv Detail & Related papers (2021-11-28T19:27:59Z) - Robust Compressed Sensing MRI with Deep Generative Priors [84.69062247243953]
We present the first successful application of the CSGM framework on clinical MRI data.
We train a generative prior on brain scans from the fastMRI dataset, and show that posterior sampling via Langevin dynamics achieves high quality reconstructions.
arXiv Detail & Related papers (2021-08-03T08:52:06Z) - The Causal Neural Connection: Expressiveness, Learnability, and
Inference [125.57815987218756]
An object called structural causal model (SCM) represents a collection of mechanisms and sources of random variation of the system under investigation.
In this paper, we show that the causal hierarchy theorem (Thm. 1, Bareinboim et al., 2020) still holds for neural models.
We introduce a special type of SCM called a neural causal model (NCM), and formalize a new type of inductive bias to encode structural constraints necessary for performing causal inferences.
arXiv Detail & Related papers (2021-07-02T01:55:18Z) - Modeling Shared Responses in Neuroimaging Studies through MultiView ICA [94.31804763196116]
Group studies involving large cohorts of subjects are important to draw general conclusions about brain functional organization.
We propose a novel MultiView Independent Component Analysis model for group studies, where data from each subject are modeled as a linear combination of shared independent sources plus noise.
We demonstrate the usefulness of our approach first on fMRI data, where our model demonstrates improved sensitivity in identifying common sources among subjects.
arXiv Detail & Related papers (2020-06-11T17:29:53Z) - Autoencoders for Unsupervised Anomaly Segmentation in Brain MR Images: A
Comparative Study [43.26668942258135]
New approaches in the field of Unsupervised Anomaly Detection (UAD) in brain MRI.
Main principle behind these works is to learn a model of normal anatomy by learning to compress and recover healthy data.
concept is of great interest to the medical image analysis community as it i) relieves from the need of vast amounts of manually segmented training data.
arXiv Detail & Related papers (2020-04-07T11:12:07Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.