Heteroscedastic Causal Structure Learning
- URL: http://arxiv.org/abs/2307.07973v1
- Date: Sun, 16 Jul 2023 07:53:16 GMT
- Title: Heteroscedastic Causal Structure Learning
- Authors: Bao Duong and Thin Nguyen
- Abstract summary: We tackle the heteroscedastic causal structure learning problem under Gaussian noises.
By exploiting the normality of the causal mechanisms, we can recover a valid causal ordering.
The result is HOST (Heteroscedastic causal STructure learning), a simple yet effective causal structure learning algorithm.
- Score: 2.566492438263125
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Heretofore, learning the directed acyclic graphs (DAGs) that encode the
cause-effect relationships embedded in observational data is a computationally
challenging problem. A recent trend of studies has shown that it is possible to
recover the DAGs with polynomial time complexity under the equal variances
assumption. However, this prohibits the heteroscedasticity of the noise, which
allows for more flexible modeling capabilities, but at the same time is
substantially more challenging to handle. In this study, we tackle the
heteroscedastic causal structure learning problem under Gaussian noises. By
exploiting the normality of the causal mechanisms, we can recover a valid
causal ordering, which can uniquely identify the causal DAG using a series of
conditional independence tests. The result is HOST (Heteroscedastic causal
STructure learning), a simple yet effective causal structure learning algorithm
that scales polynomially in both sample size and dimensionality. In addition,
via extensive empirical evaluations on a wide range of both controlled and real
datasets, we show that the proposed HOST method is competitive with
state-of-the-art approaches in both the causal order learning and structure
learning problems.
Related papers
- Effective Causal Discovery under Identifiable Heteroscedastic Noise Model [45.98718860540588]
Causal DAG learning has recently achieved promising performance in terms of both accuracy and efficiency.
We propose a novel formulation for DAG learning that accounts for the variation in noise variance across variables and observations.
We then propose an effective two-phase iterative DAG learning algorithm to address the increasing optimization difficulties.
arXiv Detail & Related papers (2023-12-20T08:51:58Z) - Causal disentanglement of multimodal data [1.589226862328831]
We introduce a causal representation learning algorithm (causalPIMA) that can use multimodal data and known physics to discover important features with causal relationships.
Our results demonstrate the capability of learning an interpretable causal structure while simultaneously discovering key features in a fully unsupervised setting.
arXiv Detail & Related papers (2023-10-27T20:30:11Z) - Causality Learning With Wasserstein Generative Adversarial Networks [2.492300648514129]
A model named DAG-WGAN combines the Wasserstein-based adversarial loss with an acyclicity constraint in an auto-encoder architecture.
It simultaneously learns causal structures while improving its data generation capability.
We compare the performance of DAG-WGAN with other models that do not involve the Wasserstein metric in order to identify its contribution to causal structure learning.
arXiv Detail & Related papers (2022-06-03T10:45:47Z) - Amortized Inference for Causal Structure Learning [72.84105256353801]
Learning causal structure poses a search problem that typically involves evaluating structures using a score or independence test.
We train a variational inference model to predict the causal structure from observational/interventional data.
Our models exhibit robust generalization capabilities under substantial distribution shift.
arXiv Detail & Related papers (2022-05-25T17:37:08Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Consistency of mechanistic causal discovery in continuous-time using
Neural ODEs [85.7910042199734]
We consider causal discovery in continuous-time for the study of dynamical systems.
We propose a causal discovery algorithm based on penalized Neural ODEs.
arXiv Detail & Related papers (2021-05-06T08:48:02Z) - Deconfounded Score Method: Scoring DAGs with Dense Unobserved
Confounding [101.35070661471124]
We show that unobserved confounding leaves a characteristic footprint in the observed data distribution that allows for disentangling spurious and causal effects.
We propose an adjusted score-based causal discovery algorithm that may be implemented with general-purpose solvers and scales to high-dimensional problems.
arXiv Detail & Related papers (2021-03-28T11:07:59Z) - Disentangling Observed Causal Effects from Latent Confounders using
Method of Moments [67.27068846108047]
We provide guarantees on identifiability and learnability under mild assumptions.
We develop efficient algorithms based on coupled tensor decomposition with linear constraints to obtain scalable and guaranteed solutions.
arXiv Detail & Related papers (2021-01-17T07:48:45Z) - A Critical View of the Structural Causal Model [89.43277111586258]
We show that one can identify the cause and the effect without considering their interaction at all.
We propose a new adversarial training method that mimics the disentangled structure of the causal model.
Our multidimensional method outperforms the literature methods on both synthetic and real world datasets.
arXiv Detail & Related papers (2020-02-23T22:52:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.