Learning Latent Causal Structures with a Redundant Input Neural Network
- URL: http://arxiv.org/abs/2003.13135v3
- Date: Tue, 8 Sep 2020 16:31:51 GMT
- Title: Learning Latent Causal Structures with a Redundant Input Neural Network
- Authors: Jonathan D. Young, Bryan Andrews, Gregory F. Cooper, Xinghua Lu
- Abstract summary: It is known that inputs cause outputs, and these causal relationships are encoded by a causal network among a set of latent variables.
We develop a deep learning model, which we call a redundant input neural network (RINN), with a modified architecture and a regularized objective function.
A series of simulation experiments provide support that the RINN method can successfully recover latent causal structure between input and output variables.
- Score: 9.044150926401574
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most causal discovery algorithms find causal structure among a set of
observed variables. Learning the causal structure among latent variables
remains an important open problem, particularly when using high-dimensional
data. In this paper, we address a problem for which it is known that inputs
cause outputs, and these causal relationships are encoded by a causal network
among a set of an unknown number of latent variables. We developed a deep
learning model, which we call a redundant input neural network (RINN), with a
modified architecture and a regularized objective function to find causal
relationships between input, hidden, and output variables. More specifically,
our model allows input variables to directly interact with all latent variables
in a neural network to influence what information the latent variables should
encode in order to generate the output variables accurately. In this setting,
the direct connections between input and latent variables makes the latent
variables partially interpretable; furthermore, the connectivity among the
latent variables in the neural network serves to model their potential causal
relationships to each other and to the output variables. A series of simulation
experiments provide support that the RINN method can successfully recover
latent causal structure between input and output variables.
Related papers
- Targeted Cause Discovery with Data-Driven Learning [66.86881771339145]
We propose a novel machine learning approach for inferring causal variables of a target variable from observations.
We employ a neural network trained to identify causality through supervised learning on simulated data.
Empirical results demonstrate the effectiveness of our method in identifying causal relationships within large-scale gene regulatory networks.
arXiv Detail & Related papers (2024-08-29T02:21:11Z) - iSCAN: Identifying Causal Mechanism Shifts among Nonlinear Additive
Noise Models [48.33685559041322]
This paper focuses on identifying the causal mechanism shifts in two or more related datasets over the same set of variables.
Code implementing the proposed method is open-source and publicly available at https://github.com/kevinsbello/iSCAN.
arXiv Detail & Related papers (2023-06-30T01:48:11Z) - BISCUIT: Causal Representation Learning from Binary Interactions [36.358968799947924]
BISCUIT is a method for simultaneously learning causal variables and their corresponding binary interaction variables.
On three robotic-inspired datasets, BISCUIT accurately identifies causal variables and can even be scaled to complex, realistic environments for embodied AI.
arXiv Detail & Related papers (2023-06-16T06:10:55Z) - Posterior Collapse and Latent Variable Non-identifiability [54.842098835445]
We propose a class of latent-identifiable variational autoencoders, deep generative models which enforce identifiability without sacrificing flexibility.
Across synthetic and real datasets, latent-identifiable variational autoencoders outperform existing methods in mitigating posterior collapse and providing meaningful representations of the data.
arXiv Detail & Related papers (2023-01-02T06:16:56Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Systematic Evaluation of Causal Discovery in Visual Model Based
Reinforcement Learning [76.00395335702572]
A central goal for AI and causality is the joint discovery of abstract representations and causal structure.
Existing environments for studying causal induction are poorly suited for this objective because they have complicated task-specific causal graphs.
In this work, our goal is to facilitate research in learning representations of high-level variables as well as causal structures among them.
arXiv Detail & Related papers (2021-07-02T05:44:56Z) - Learning latent causal graphs via mixture oracles [40.71943453524747]
We study the problem of reconstructing a causal graphical model from data in the presence of latent variables.
The main problem of interest is recovering the causal structure over the latent variables while allowing for general, potentially nonlinear dependence between the variables.
arXiv Detail & Related papers (2021-06-29T16:53:34Z) - Disentangling Observed Causal Effects from Latent Confounders using
Method of Moments [67.27068846108047]
We provide guarantees on identifiability and learnability under mild assumptions.
We develop efficient algorithms based on coupled tensor decomposition with linear constraints to obtain scalable and guaranteed solutions.
arXiv Detail & Related papers (2021-01-17T07:48:45Z) - Visual Neural Decomposition to Explain Multivariate Data Sets [13.117139248511783]
Investigating relationships between variables in multi-dimensional data sets is a common task for data analysts and engineers.
We propose a novel approach to visualize correlations between input variables and a target output variable that scales to hundreds of variables.
arXiv Detail & Related papers (2020-09-11T15:53:37Z) - Multi-Task Learning for Multi-Dimensional Regression: Application to
Luminescence Sensing [0.0]
A new approach to non-linear regression is to use neural networks, particularly feed-forward architectures with a sufficient number of hidden layers and an appropriate number of output neurons.
We propose multi-task learning (MTL) architectures. These are characterized by multiple branches of task-specific layers, which have as input the output of a common set of layers.
To demonstrate the power of this approach for multi-dimensional regression, the method is applied to luminescence sensing.
arXiv Detail & Related papers (2020-07-27T21:23:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.