Causal disentanglement of multimodal data
- URL: http://arxiv.org/abs/2310.18471v2
- Date: Wed, 8 Nov 2023 18:54:52 GMT
- Title: Causal disentanglement of multimodal data
- Authors: Elise Walker, Jonas A. Actor, Carianne Martinez, and Nathaniel Trask
- Abstract summary: We introduce a causal representation learning algorithm (causalPIMA) that can use multimodal data and known physics to discover important features with causal relationships.
Our results demonstrate the capability of learning an interpretable causal structure while simultaneously discovering key features in a fully unsupervised setting.
- Score: 1.589226862328831
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Causal representation learning algorithms discover lower-dimensional
representations of data that admit a decipherable interpretation of cause and
effect; as achieving such interpretable representations is challenging, many
causal learning algorithms utilize elements indicating prior information, such
as (linear) structural causal models, interventional data, or weak supervision.
Unfortunately, in exploratory causal representation learning, such elements and
prior information may not be available or warranted. Alternatively, scientific
datasets often have multiple modalities or physics-based constraints, and the
use of such scientific, multimodal data has been shown to improve
disentanglement in fully unsupervised settings. Consequently, we introduce a
causal representation learning algorithm (causalPIMA) that can use multimodal
data and known physics to discover important features with causal
relationships. Our innovative algorithm utilizes a new differentiable
parametrization to learn a directed acyclic graph (DAG) together with a latent
space of a variational autoencoder in an end-to-end differentiable framework
via a single, tractable evidence lower bound loss function. We place a Gaussian
mixture prior on the latent space and identify each of the mixtures with an
outcome of the DAG nodes; this novel identification enables feature discovery
with causal relationships. Tested against a synthetic and a scientific dataset,
our results demonstrate the capability of learning an interpretable causal
structure while simultaneously discovering key features in a fully unsupervised
setting.
Related papers
- Learning to refine domain knowledge for biological network inference [2.209921757303168]
Perturbation experiments allow biologists to discover causal relationships between variables of interest.
The sparsity and high dimensionality of these data pose significant challenges for causal structure learning algorithms.
We propose an amortized algorithm for refining domain knowledge, based on data observations.
arXiv Detail & Related papers (2024-10-18T12:53:23Z) - Multi-modal Causal Structure Learning and Root Cause Analysis [67.67578590390907]
We propose Mulan, a unified multi-modal causal structure learning method for root cause localization.
We leverage a log-tailored language model to facilitate log representation learning, converting log sequences into time-series data.
We also introduce a novel key performance indicator-aware attention mechanism for assessing modality reliability and co-learning a final causal graph.
arXiv Detail & Related papers (2024-02-04T05:50:38Z) - Multi-granularity Causal Structure Learning [23.125497987255237]
We develop MgCSL (Multi-granularity Causal Structure Learning), which first leverages sparse auto-encoder to explore coarse-graining strategies and causal abstractions.
MgCSL then takes multi-granularity variables as inputs to train multilayer perceptrons and to delve the causality between variables.
Experimental results show that MgCSL outperforms competitive baselines, and finds out explainable causal connections on fMRI datasets.
arXiv Detail & Related papers (2023-12-09T11:35:25Z) - Heteroscedastic Causal Structure Learning [2.566492438263125]
We tackle the heteroscedastic causal structure learning problem under Gaussian noises.
By exploiting the normality of the causal mechanisms, we can recover a valid causal ordering.
The result is HOST (Heteroscedastic causal STructure learning), a simple yet effective causal structure learning algorithm.
arXiv Detail & Related papers (2023-07-16T07:53:16Z) - Amortized Inference for Causal Structure Learning [72.84105256353801]
Learning causal structure poses a search problem that typically involves evaluating structures using a score or independence test.
We train a variational inference model to predict the causal structure from observational/interventional data.
Our models exhibit robust generalization capabilities under substantial distribution shift.
arXiv Detail & Related papers (2022-05-25T17:37:08Z) - Equivariance Allows Handling Multiple Nuisance Variables When Analyzing
Pooled Neuroimaging Datasets [53.34152466646884]
In this paper, we show how bringing recent results on equivariant representation learning instantiated on structured spaces together with simple use of classical results on causal inference provides an effective practical solution.
We demonstrate how our model allows dealing with more than one nuisance variable under some assumptions and can enable analysis of pooled scientific datasets in scenarios that would otherwise entail removing a large portion of the samples.
arXiv Detail & Related papers (2022-03-29T04:54:06Z) - Learning Neural Causal Models with Active Interventions [83.44636110899742]
We introduce an active intervention-targeting mechanism which enables a quick identification of the underlying causal structure of the data-generating process.
Our method significantly reduces the required number of interactions compared with random intervention targeting.
We demonstrate superior performance on multiple benchmarks from simulated to real-world data.
arXiv Detail & Related papers (2021-09-06T13:10:37Z) - The Causal Neural Connection: Expressiveness, Learnability, and
Inference [125.57815987218756]
An object called structural causal model (SCM) represents a collection of mechanisms and sources of random variation of the system under investigation.
In this paper, we show that the causal hierarchy theorem (Thm. 1, Bareinboim et al., 2020) still holds for neural models.
We introduce a special type of SCM called a neural causal model (NCM), and formalize a new type of inductive bias to encode structural constraints necessary for performing causal inferences.
arXiv Detail & Related papers (2021-07-02T01:55:18Z) - Consistency of mechanistic causal discovery in continuous-time using
Neural ODEs [85.7910042199734]
We consider causal discovery in continuous-time for the study of dynamical systems.
We propose a causal discovery algorithm based on penalized Neural ODEs.
arXiv Detail & Related papers (2021-05-06T08:48:02Z) - Causal Discovery from Incomplete Data: A Deep Learning Approach [21.289342482087267]
Imputated Causal Learning is proposed to perform iterative missing data imputation and causal structure discovery.
We show that ICL can outperform state-of-the-art methods under different missing data mechanisms.
arXiv Detail & Related papers (2020-01-15T14:28:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.