Learning Cyclic Causal Models from Incomplete Data
- URL: http://arxiv.org/abs/2402.15625v1
- Date: Fri, 23 Feb 2024 22:03:12 GMT
- Title: Learning Cyclic Causal Models from Incomplete Data
- Authors: Muralikrishnna G. Sethuraman, Faramarz Fekri
- Abstract summary: We propose a novel framework, named MissNODAGS, for learning cyclic causal graphs from partially missing data.
Under the additive noise model, MissNODAGS learns the causal graph by alternating between imputing the missing data and maximizing the expected log-likelihood of the visible part of the data.
We demonstrate improved performance when compared to using state-of-the-art imputation techniques followed by causal learning on partially missing interventional data.
- Score: 13.69726643902085
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Causal learning is a fundamental problem in statistics and science, offering
insights into predicting the effects of unseen treatments on a system. Despite
recent advances in this topic, most existing causal discovery algorithms
operate under two key assumptions: (i) the underlying graph is acyclic, and
(ii) the available data is complete. These assumptions can be problematic as
many real-world systems contain feedback loops (e.g., biological systems), and
practical scenarios frequently involve missing data. In this work, we propose a
novel framework, named MissNODAGS, for learning cyclic causal graphs from
partially missing data. Under the additive noise model, MissNODAGS learns the
causal graph by alternating between imputing the missing data and maximizing
the expected log-likelihood of the visible part of the data in each training
step, following the principles of the expectation-maximization (EM) framework.
Through synthetic experiments and real-world single-cell perturbation data, we
demonstrate improved performance when compared to using state-of-the-art
imputation techniques followed by causal learning on partially missing
interventional data.
Related papers
- MissNODAG: Differentiable Cyclic Causal Graph Learning from Incomplete Data [13.006241141102]
We propose MissNODAG, a framework for learning the underlying cyclic causal graph and the missingness mechanism from partially observed data.
Our framework integrates an additive noise model with an expectation-maximization procedure, alternating between imputing missing values and optimizing the observed data likelihood.
We demonstrate the effectiveness of MissNODAG through synthetic experiments and an application to real-world gene perturbation data.
arXiv Detail & Related papers (2024-10-24T17:09:10Z) - A Comparative Study of Machine Learning Models Predicting Energetics of Interacting Defects [5.574191640970887]
We present a comparative study of three different methods to predict the free energy change of systems with interacting defects.
Our findings indicate that the cluster expansion model can achieve precise energetics predictions even with this limited dataset.
This research provide a preliminary evaluation of applying machine learning techniques in imperfect surface systems.
arXiv Detail & Related papers (2024-03-20T02:15:48Z) - Multi-modal Causal Structure Learning and Root Cause Analysis [67.67578590390907]
We propose Mulan, a unified multi-modal causal structure learning method for root cause localization.
We leverage a log-tailored language model to facilitate log representation learning, converting log sequences into time-series data.
We also introduce a novel key performance indicator-aware attention mechanism for assessing modality reliability and co-learning a final causal graph.
arXiv Detail & Related papers (2024-02-04T05:50:38Z) - Causal disentanglement of multimodal data [1.589226862328831]
We introduce a causal representation learning algorithm (causalPIMA) that can use multimodal data and known physics to discover important features with causal relationships.
Our results demonstrate the capability of learning an interpretable causal structure while simultaneously discovering key features in a fully unsupervised setting.
arXiv Detail & Related papers (2023-10-27T20:30:11Z) - NODAGS-Flow: Nonlinear Cyclic Causal Structure Learning [8.20217860574125]
We propose a novel framework for learning nonlinear cyclic causal models from interventional data, called NODAGS-Flow.
We show significant performance improvements with our approach compared to state-of-the-art methods with respect to structure recovery and predictive performance.
arXiv Detail & Related papers (2023-01-04T23:28:18Z) - MissDAG: Causal Discovery in the Presence of Missing Data with
Continuous Additive Noise Models [78.72682320019737]
We develop a general method, which we call MissDAG, to perform causal discovery from data with incomplete observations.
MissDAG maximizes the expected likelihood of the visible part of observations under the expectation-maximization framework.
We demonstrate the flexibility of MissDAG for incorporating various causal discovery algorithms and its efficacy through extensive simulations and real data experiments.
arXiv Detail & Related papers (2022-05-27T09:59:46Z) - Predicting Seriousness of Injury in a Traffic Accident: A New Imbalanced
Dataset and Benchmark [62.997667081978825]
The paper introduces a new dataset to assess the performance of machine learning algorithms in the prediction of the seriousness of injury in a traffic accident.
The dataset is created by aggregating publicly available datasets from the UK Department for Transport.
arXiv Detail & Related papers (2022-05-20T21:15:26Z) - MIRACLE: Causally-Aware Imputation via Learning Missing Data Mechanisms [82.90843777097606]
We propose a causally-aware imputation algorithm (MIRACLE) for missing data.
MIRACLE iteratively refines the imputation of a baseline by simultaneously modeling the missingness generating mechanism.
We conduct extensive experiments on synthetic and a variety of publicly available datasets to show that MIRACLE is able to consistently improve imputation.
arXiv Detail & Related papers (2021-11-04T22:38:18Z) - Distributionally Robust Semi-Supervised Learning Over Graphs [68.29280230284712]
Semi-supervised learning (SSL) over graph-structured data emerges in many network science applications.
To efficiently manage learning over graphs, variants of graph neural networks (GNNs) have been developed recently.
Despite their success in practice, most of existing methods are unable to handle graphs with uncertain nodal attributes.
Challenges also arise due to distributional uncertainties associated with data acquired by noisy measurements.
A distributionally robust learning framework is developed, where the objective is to train models that exhibit quantifiable robustness against perturbations.
arXiv Detail & Related papers (2021-10-20T14:23:54Z) - Learning Neural Causal Models with Active Interventions [83.44636110899742]
We introduce an active intervention-targeting mechanism which enables a quick identification of the underlying causal structure of the data-generating process.
Our method significantly reduces the required number of interactions compared with random intervention targeting.
We demonstrate superior performance on multiple benchmarks from simulated to real-world data.
arXiv Detail & Related papers (2021-09-06T13:10:37Z) - Causal Discovery from Incomplete Data: A Deep Learning Approach [21.289342482087267]
Imputated Causal Learning is proposed to perform iterative missing data imputation and causal structure discovery.
We show that ICL can outperform state-of-the-art methods under different missing data mechanisms.
arXiv Detail & Related papers (2020-01-15T14:28:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.