Identifying Causal Direction via Dense Functional Classes
- URL: http://arxiv.org/abs/2509.00538v1
- Date: Sat, 30 Aug 2025 15:42:31 GMT
- Title: Identifying Causal Direction via Dense Functional Classes
- Authors: Katerina Hlavackova-Schindler, Suzana Marsela,
- Abstract summary: We show that LCUBE achieves superior precision in terms of AUDRC on the real-world Tuebingen cause-effect pairs dataset.<n>It also shows superior average precision across common 10 benchmark datasets and achieves above average precision on 13 datasets.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We address the problem of determining the causal direction between two univariate, continuous-valued variables, X and Y, under the assumption of no hidden confounders. In general, it is not possible to make definitive statements about causality without some assumptions on the underlying model. To distinguish between cause and effect, we propose a bivariate causal score based on the Minimum Description Length (MDL) principle, using functions that possess the density property on a compact real interval. We prove the identifiability of these causal scores under specific conditions. These conditions can be easily tested. Gaussianity of the noise in the causal model equations is not assumed, only that the noise is low. The well-studied class of cubic splines possesses the density property on a compact real interval. We propose LCUBE as an instantiation of the MDL-based causal score utilizing cubic regression splines. LCUBE is an identifiable method that is also interpretable, simple, and very fast. It has only one hyperparameter. Empirical evaluations compared to state-of-the-art methods demonstrate that LCUBE achieves superior precision in terms of AUDRC on the real-world Tuebingen cause-effect pairs dataset. It also shows superior average precision across common 10 benchmark datasets and achieves above average precision on 13 datasets.
Related papers
- Distinguishing Cause from Effect with Causal Velocity Models [3.0523869645673076]
We develop a method for causal discovery that extends beyond known model classes such as additive or location scale noise.<n>When the score is estimated well, the objective is also useful for detecting model non-identifiability and misspecification.
arXiv Detail & Related papers (2025-02-07T17:50:14Z) - Invariant Causal Prediction with Local Models [52.161513027831646]
We consider the task of identifying the causal parents of a target variable among a set of candidates from observational data.
We introduce a practical method called L-ICP ($textbfL$ocalized $textbfI$nvariant $textbfCa$usal $textbfP$rediction), which is based on a hypothesis test for parent identification using a ratio of minimum and maximum statistics.
arXiv Detail & Related papers (2024-01-10T15:34:42Z) - Effective Causal Discovery under Identifiable Heteroscedastic Noise Model [45.98718860540588]
Causal DAG learning has recently achieved promising performance in terms of both accuracy and efficiency.
We propose a novel formulation for DAG learning that accounts for the variation in noise variance across variables and observations.
We then propose an effective two-phase iterative DAG learning algorithm to address the increasing optimization difficulties.
arXiv Detail & Related papers (2023-12-20T08:51:58Z) - Structural restrictions in local causal discovery: identifying direct causes of a target variable [0.9208007322096533]
Learning a set of direct causes of a target variable from an observational joint distribution is a fundamental problem in science.<n>Here, we are only interested in identifying the direct causes of one target variable, not the full DAG.<n>This allows us to relax the identifiability assumptions and develop possibly faster and more robust algorithms.
arXiv Detail & Related papers (2023-07-29T18:31:35Z) - A Scale-Invariant Sorting Criterion to Find a Causal Order in Additive
Noise Models [49.038420266408586]
We show that sorting variables by increasing variance often yields an ordering close to a causal order.
We propose an efficient baseline algorithm termed $R2$-SortnRegress that exploits high $R2$-sortability.
Our findings reveal high $R2$-sortability as an assumption about the data generating process relevant to causal discovery.
arXiv Detail & Related papers (2023-03-31T17:05:46Z) - Distinguishing Cause from Effect on Categorical Data: The Uniform
Channel Model [0.0]
Distinguishing cause from effect using observations of a pair of random variables is a core problem in causal discovery.
We propose a criterion to address the cause-effect problem with categorical variables.
We select as the most likely causal direction the one in which the conditional probability mass function is closer to a uniform channel (UC)
arXiv Detail & Related papers (2023-03-14T13:54:11Z) - Cause-Effect Inference in Location-Scale Noise Models: Maximum
Likelihood vs. Independence Testing [19.23479356810746]
A fundamental problem of causal discovery is cause-effect inference, learning the correct causal direction between two random variables.
Recently introduced heteroscedastic location-scale noise functional models (LSNMs) combine expressive power with identifiability guarantees.
We show that LSNM model selection based on maximizing likelihood achieves state-of-the-art accuracy, when the noise distributions are correctly specified.
arXiv Detail & Related papers (2023-01-26T20:48:32Z) - Optimizing the Noise in Self-Supervised Learning: from Importance
Sampling to Noise-Contrastive Estimation [80.07065346699005]
It is widely assumed that the optimal noise distribution should be made equal to the data distribution, as in Generative Adversarial Networks (GANs)
We turn to Noise-Contrastive Estimation which grounds this self-supervised task as an estimation problem of an energy-based model of the data.
We soberly conclude that the optimal noise may be hard to sample from, and the gain in efficiency can be modest compared to choosing the noise distribution equal to the data's.
arXiv Detail & Related papers (2023-01-23T19:57:58Z) - The Optimal Noise in Noise-Contrastive Learning Is Not What You Think [80.07065346699005]
We show that deviating from this assumption can actually lead to better statistical estimators.
In particular, the optimal noise distribution is different from the data's and even from a different family.
arXiv Detail & Related papers (2022-03-02T13:59:20Z) - Partial Identification with Noisy Covariates: A Robust Optimization
Approach [94.10051154390237]
Causal inference from observational datasets often relies on measuring and adjusting for covariates.
We show that this robust optimization approach can extend a wide range of causal adjustment methods to perform partial identification.
Across synthetic and real datasets, we find that this approach provides ATE bounds with a higher coverage probability than existing methods.
arXiv Detail & Related papers (2022-02-22T04:24:26Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Causal Order Identification to Address Confounding: Binary Variables [4.56877715768796]
This paper considers an extension of the linear non-Gaussian acyclic model (LiNGAM)
LiNGAM determines the causal order among variables from a dataset when the variables are expressed by a set of linear equations, including noise.
arXiv Detail & Related papers (2021-08-10T22:09:43Z) - Causal Expectation-Maximisation [70.45873402967297]
We show that causal inference is NP-hard even in models characterised by polytree-shaped graphs.
We introduce the causal EM algorithm to reconstruct the uncertainty about the latent variables from data about categorical manifest variables.
We argue that there appears to be an unnoticed limitation to the trending idea that counterfactual bounds can often be computed without knowledge of the structural equations.
arXiv Detail & Related papers (2020-11-04T10:25:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.