Learning Causal Graphs via Monotone Triangular Transport Maps
- URL: http://arxiv.org/abs/2305.18210v1
- Date: Fri, 26 May 2023 13:24:17 GMT
- Title: Learning Causal Graphs via Monotone Triangular Transport Maps
- Authors: Sina Akbari, Luca Ganassali
- Abstract summary: We study the problem of causal structure learning from data using optimal transport (OT)
We provide an algorithm for causal discovery up to Markov Equivalence with no assumptions on the structural equations/noise distributions.
We provide experimental results to compare the proposed approach with the state of the art on both synthetic and real-world datasets.
- Score: 1.6752182911522522
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of causal structure learning from data using optimal
transport (OT). Specifically, we first provide a constraint-based method which
builds upon lower-triangular monotone parametric transport maps to design
conditional independence tests which are agnostic to the noise distribution. We
provide an algorithm for causal discovery up to Markov Equivalence with no
assumptions on the structural equations/noise distributions, which allows for
settings with latent variables. Our approach also extends to score-based causal
discovery by providing a novel means for defining scores. This allows us to
uniquely recover the causal graph under additional identifiability and
structural assumptions, such as additive noise or post-nonlinear models. We
provide experimental results to compare the proposed approach with the state of
the art on both synthetic and real-world datasets.
Related papers
- Induced Covariance for Causal Discovery in Linear Sparse Structures [55.2480439325792]
Causal models seek to unravel the cause-effect relationships among variables from observed data.
This paper introduces a novel causal discovery algorithm designed for settings in which variables exhibit linearly sparse relationships.
arXiv Detail & Related papers (2024-10-02T04:01:38Z) - Sub-graph Based Diffusion Model for Link Prediction [43.15741675617231]
Denoising Diffusion Probabilistic Models (DDPMs) represent a contemporary class of generative models with exceptional qualities.
We build a novel generative model for link prediction using a dedicated design to decompose the likelihood estimation process via the Bayesian formula.
Our proposed method presents numerous advantages: (1) transferability across datasets without retraining, (2) promising generalization on limited training data, and (3) robustness against graph adversarial attacks.
arXiv Detail & Related papers (2024-09-13T02:23:55Z) - Hybrid Top-Down Global Causal Discovery with Local Search for Linear and Nonlinear Additive Noise Models [2.0738462952016232]
Methods based on functional causal models can identify a unique graph, but suffer from the curse of dimensionality or impose strong parametric assumptions.
We propose a novel hybrid approach for global causal discovery in observational data that leverages local causal substructures.
We provide theoretical guarantees for correctness and worst-case time complexities, with empirical validation on synthetic data.
arXiv Detail & Related papers (2024-05-23T12:28:16Z) - Stochastic interpolants with data-dependent couplings [31.854717378556334]
We use the framework of interpolants to formalize how to itcouple the base and the target densities.
We show that these transport maps can be learned by solving a simple square loss regression problem analogous to the standard independent setting.
arXiv Detail & Related papers (2023-10-05T17:46:31Z) - Causal Discovery with Score Matching on Additive Models with Arbitrary
Noise [37.13308785728276]
Causal discovery methods are intrinsically constrained by the set of assumptions needed to ensure structure identifiability.
In this paper we show the shortcomings of inference under this hypothesis, analyzing the risk of edge inversion under violation of Gaussianity of the noise terms.
We propose a novel method for inferring the topological ordering of the variables in the causal graph, from data generated according to an additive non-linear model with a generic noise distribution.
This leads to NoGAM, a causal discovery algorithm with a minimal set of assumptions and state of the art performance, experimentally benchmarked on synthetic data.
arXiv Detail & Related papers (2023-04-06T17:50:46Z) - Amortized Inference for Causal Structure Learning [72.84105256353801]
Learning causal structure poses a search problem that typically involves evaluating structures using a score or independence test.
We train a variational inference model to predict the causal structure from observational/interventional data.
Our models exhibit robust generalization capabilities under substantial distribution shift.
arXiv Detail & Related papers (2022-05-25T17:37:08Z) - Score matching enables causal discovery of nonlinear additive noise
models [63.93669924730725]
We show how to design a new generation of scalable causal discovery methods.
We propose a new efficient method for approximating the score's Jacobian, enabling to recover the causal graph.
arXiv Detail & Related papers (2022-03-08T21:34:46Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Learning non-Gaussian graphical models via Hessian scores and triangular
transport [6.308539010172309]
We propose an algorithm for learning the Markov structure of continuous and non-Gaussian distributions.
Our algorithm SING estimates the density using a deterministic coupling, induced by a triangular transport map, and iteratively exploits sparse structure in the map to reveal sparsity in the graph.
arXiv Detail & Related papers (2021-01-08T16:42:42Z) - Learning Noise-Aware Encoder-Decoder from Noisy Labels by Alternating
Back-Propagation for Saliency Detection [54.98042023365694]
We propose a noise-aware encoder-decoder framework to disentangle a clean saliency predictor from noisy training examples.
The proposed model consists of two sub-models parameterized by neural networks.
arXiv Detail & Related papers (2020-07-23T18:47:36Z) - Generative Modeling with Denoising Auto-Encoders and Langevin Sampling [88.83704353627554]
We show that both DAE and DSM provide estimates of the score of the smoothed population density.
We then apply our results to the homotopy method of arXiv:1907.05600 and provide theoretical justification for its empirical success.
arXiv Detail & Related papers (2020-01-31T23:50:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.