Improving the Efficiency of the PC Algorithm by Using Model-Based
Conditional Independence Tests
- URL: http://arxiv.org/abs/2211.06536v1
- Date: Sat, 12 Nov 2022 00:54:53 GMT
- Title: Improving the Efficiency of the PC Algorithm by Using Model-Based
Conditional Independence Tests
- Authors: Erica Cai, Andrew McGregor, David Jensen
- Abstract summary: Constraint-based structure learning algorithms such as PC use conditional independence (CI) tests to infer causal structure.
Traditionally, constraint-based algorithms perform CI tests with a preference for smaller-sized conditioning sets.
We propose a new strategy for constraint-based algorithms which may result in a reduction of the total number of CI tests performed.
- Score: 11.33674514817655
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning causal structure is useful in many areas of artificial intelligence,
including planning, robotics, and explanation. Constraint-based structure
learning algorithms such as PC use conditional independence (CI) tests to infer
causal structure. Traditionally, constraint-based algorithms perform CI tests
with a preference for smaller-sized conditioning sets, partially because the
statistical power of conventional CI tests declines rapidly as the size of the
conditioning set increases. However, many modern conditional independence tests
are model-based, and these tests use well-regularized models that maintain
statistical power even with very large conditioning sets. This suggests an
intriguing new strategy for constraint-based algorithms which may result in a
reduction of the total number of CI tests performed: Test variable pairs with
large conditioning sets first, as a pre-processing step that finds some
conditional independencies quickly, before moving on to the more conventional
strategy that favors small conditioning sets. We propose such a pre-processing
step for the PC algorithm which relies on performing CI tests on a few randomly
selected large conditioning sets. We perform an empirical analysis on directed
acyclic graphs (DAGs) that correspond to real-world systems and both empirical
and theoretical analyses for Erd\H{o}s-Renyi DAGs. Our results show that
Pre-Processing Plus PC (P3PC) performs far fewer CI tests than the original PC
algorithm, between 0.5% to 36%, and often less than 10%, of the CI tests that
the PC algorithm alone performs. The efficiency gains are particularly
significant for the DAGs corresponding to real-world systems.
Related papers
- Fast Flow Matching based Conditional Independence Tests for Causal Discovery [19.33167245211968]
Constraint-based causal discovery methods require a large number of conditional independence (CI) tests.<n>We propose the Flow Matching-based Conditional Independence Test (FMCIT)<n>The proposed test leverages the high computational efficiency of flow matching and requires the model to be trained only once throughout the entire causal discovery procedure.
arXiv Detail & Related papers (2026-02-09T06:43:23Z) - Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs [76.43407125275202]
o1-like models can emulate human-like long-time thinking during inference.
This paper presents the first comprehensive study on the prevalent issue of overthinking in these models.
We propose strategies to mitigate overthinking, streamlining reasoning processes without compromising accuracy.
arXiv Detail & Related papers (2024-12-30T18:55:12Z) - Testing Causal Models with Hidden Variables in Polynomial Delay via Conditional Independencies [49.99600569996907]
Testing a hypothesized causal model against observational data is a key prerequisite for many causal inference tasks.
While a model can assume exponentially many conditional independence relations (CIs), testing all of them is both impractical and unnecessary.
We introduce c-LMP for causal graphs with hidden variables and develop a delay algorithm to list these CIs in poly-time intervals.
arXiv Detail & Related papers (2024-09-22T21:05:56Z) - Hyperparameters in Continual Learning: A Reality Check [53.30082523545212]
Continual learning (CL) aims to train a model on a sequence of tasks while balancing the trade-off between plasticity (learning new tasks) and stability (retaining prior knowledge)
The dominantly adopted conventional evaluation protocol for CL algorithms selects the best hyper parameters in a given scenario and then evaluates the algorithms in the same scenario.
This protocol has significant shortcomings: it overestimates the CL capacity of algorithms and relies on unrealistic hyper parameter tuning.
We argue that the evaluation of CL algorithms should focus on assessing the generalizability of their CL capacity to unseen scenarios.
arXiv Detail & Related papers (2024-03-14T03:13:01Z) - Precise Error Rates for Computationally Efficient Testing [75.63895690909241]
We revisit the question of simple-versus-simple hypothesis testing with an eye towards computational complexity.
An existing test based on linear spectral statistics achieves the best possible tradeoff curve between type I and type II error rates.
arXiv Detail & Related papers (2023-11-01T04:41:16Z) - A Stable, Fast, and Fully Automatic Learning Algorithm for Predictive
Coding Networks [65.34977803841007]
Predictive coding networks are neuroscience-inspired models with roots in both Bayesian statistics and neuroscience.
We show how by simply changing the temporal scheduling of the update rule for the synaptic weights leads to an algorithm that is much more efficient and stable than the original one.
arXiv Detail & Related papers (2022-11-16T00:11:04Z) - A Simple Unified Approach to Testing High-Dimensional Conditional
Independences for Categorical and Ordinal Data [0.26651200086513094]
Conditional independence (CI) tests underlie many approaches to model testing and structure learning in causal inference.
Most existing CI tests for categorical and ordinal data stratify the sample by the conditioning variables, perform simple independence tests in each stratum, and combine the results.
Here we propose a simple unified CI test for ordinal and categorical data that maintains reasonable calibration and power in high dimensions.
arXiv Detail & Related papers (2022-06-09T08:56:12Z) - Recursive Causal Structure Learning in the Presence of Latent Variables
and Selection Bias [27.06618125828978]
We consider the problem of learning the causal MAG of a system from observational data in the presence of latent variables and selection bias.
We propose a novel computationally efficient constraint-based method that is sound and complete.
We provide experimental results to compare the proposed approach with the state of the art on both synthetic and real-world structures.
arXiv Detail & Related papers (2021-10-22T19:49:59Z) - Recovering Causal Structures from Low-Order Conditional Independencies [6.891238879512672]
We propose an algorithm for a given set of conditional independencies of order less or equal to $k$, where $k$ is a small fixed number.
Our results complete and generalize the previous work on learning from pairwise marginal independencies.
arXiv Detail & Related papers (2020-10-06T12:47:20Z) - Statistical Query Algorithms and Low-Degree Tests Are Almost Equivalent [29.684442397855197]
We study two of the most popular restricted computational models, the statistical query framework and low-degree corollas, in the context of high-dimensional hypothesis testing.
Under mild conditions on the testing problem, the two classes of algorithms are essentially equivalent in power.
Asries, we obtain new statistical query lower bounds for sparse PCA, tensor PCA and several variants of the planted clique problem.
arXiv Detail & Related papers (2020-09-13T22:55:18Z) - A Constraint-Based Algorithm for the Structural Learning of
Continuous-Time Bayesian Networks [70.88503833248159]
We propose the first constraint-based algorithm for learning the structure of continuous-time Bayesian networks.
We discuss the different statistical tests and the underlying hypotheses used by our proposal to establish conditional independence.
arXiv Detail & Related papers (2020-07-07T07:34:09Z) - Approximation Algorithms for Sparse Principal Component Analysis [57.5357874512594]
Principal component analysis (PCA) is a widely used dimension reduction technique in machine learning and statistics.
Various approaches to obtain sparse principal direction loadings have been proposed, which are termed Sparse Principal Component Analysis.
We present thresholding as a provably accurate, time, approximation algorithm for the SPCA problem.
arXiv Detail & Related papers (2020-06-23T04:25:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.