Active Learning of Causal Structures with Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2009.03009v1
- Date: Mon, 7 Sep 2020 10:49:06 GMT
- Title: Active Learning of Causal Structures with Deep Reinforcement Learning
- Authors: Amir Amirinezhad, Saber Salehkaleybar, Matin Hashemi
- Abstract summary: We study the problem of experiment design to learn causal structures from interventional data.
We present the first deep reinforcement learning based solution for the problem of experiment design.
- Score: 13.202747831999414
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of experiment design to learn causal structures from
interventional data. We consider an active learning setting in which the
experimenter decides to intervene on one of the variables in the system in each
step and uses the results of the intervention to recover further causal
relationships among the variables. The goal is to fully identify the causal
structures with minimum number of interventions. We present the first deep
reinforcement learning based solution for the problem of experiment design. In
the proposed method, we embed input graphs to vectors using a graph neural
network and feed them to another neural network which outputs a variable for
performing intervention in each step. Both networks are trained jointly via a
Q-iteration algorithm. Experimental results show that the proposed method
achieves competitive performance in recovering causal structures with respect
to previous works, while significantly reducing execution time in dense graphs.
Related papers
- Targeted Cause Discovery with Data-Driven Learning [66.86881771339145]
We propose a novel machine learning approach for inferring causal variables of a target variable from observations.
We employ a neural network trained to identify causality through supervised learning on simulated data.
Empirical results demonstrate the effectiveness of our method in identifying causal relationships within large-scale gene regulatory networks.
arXiv Detail & Related papers (2024-08-29T02:21:11Z) - Multi-modal Causal Structure Learning and Root Cause Analysis [67.67578590390907]
We propose Mulan, a unified multi-modal causal structure learning method for root cause localization.
We leverage a log-tailored language model to facilitate log representation learning, converting log sequences into time-series data.
We also introduce a novel key performance indicator-aware attention mechanism for assessing modality reliability and co-learning a final causal graph.
arXiv Detail & Related papers (2024-02-04T05:50:38Z) - CORE: Towards Scalable and Efficient Causal Discovery with Reinforcement
Learning [2.7446241148152253]
CORE is a reinforcement learning-based approach for causal discovery and intervention planning.
Our results demonstrate that CORE generalizes to unseen graphs and efficiently uncovers causal structures.
CORE scales to larger graphs with up to 10 variables and outperforms existing approaches in structure estimation accuracy and sample efficiency.
arXiv Detail & Related papers (2024-01-30T12:57:52Z) - Learning Neural Causal Models with Active Interventions [83.44636110899742]
We introduce an active intervention-targeting mechanism which enables a quick identification of the underlying causal structure of the data-generating process.
Our method significantly reduces the required number of interactions compared with random intervention targeting.
We demonstrate superior performance on multiple benchmarks from simulated to real-world data.
arXiv Detail & Related papers (2021-09-06T13:10:37Z) - Near-Optimal Multi-Perturbation Experimental Design for Causal Structure
Learning [0.0]
Causal structures can be learnt by performing experiments on the system of interest.
We address the largely unexplored problem of designing experiments that simultaneously intervene on multiple variables.
arXiv Detail & Related papers (2021-05-28T18:00:00Z) - Multi-task Supervised Learning via Cross-learning [102.64082402388192]
We consider a problem known as multi-task learning, consisting of fitting a set of regression functions intended for solving different tasks.
In our novel formulation, we couple the parameters of these functions, so that they learn in their task specific domains while staying close to each other.
This facilitates cross-fertilization in which data collected across different domains help improving the learning performance at each other task.
arXiv Detail & Related papers (2020-10-24T21:35:57Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z) - Parallelization Techniques for Verifying Neural Networks [52.917845265248744]
We introduce an algorithm based on the verification problem in an iterative manner and explore two partitioning strategies.
We also introduce a highly parallelizable pre-processing algorithm that uses the neuron activation phases to simplify the neural network verification problems.
arXiv Detail & Related papers (2020-04-17T20:21:47Z) - Latent Network Structure Learning from High Dimensional Multivariate
Point Processes [5.079425170410857]
We propose a new class of nonstationary Hawkes processes to characterize the complex processes underlying the observed data.
We estimate the latent network structure using an efficient sparse least squares estimation approach.
We demonstrate the efficacy of our proposed method through simulation studies and an application to a neuron spike train data set.
arXiv Detail & Related papers (2020-04-07T17:48:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.