Invariant Causal Set Covering Machines
- URL: http://arxiv.org/abs/2306.04777v2
- Date: Wed, 19 Jul 2023 22:41:05 GMT
- Title: Invariant Causal Set Covering Machines
- Authors: Thibaud Godon, Baptiste Bauvin, Pascal Germain, Jacques Corbeil,
Alexandre Drouin
- Abstract summary: Rule-based models, such as decision trees, appeal to practitioners due to their interpretable nature.
However, the learning algorithms that produce such models are often vulnerable to spurious associations and thus, they are not guaranteed to extract causally-relevant insights.
We propose Invariant Causal Set Covering Machines, an extension of the classical Set Covering Machine algorithm for conjunctions/disjunctions of binary-valued rules that provably avoids spurious associations.
- Score: 64.86459157191346
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Rule-based models, such as decision trees, appeal to practitioners due to
their interpretable nature. However, the learning algorithms that produce such
models are often vulnerable to spurious associations and thus, they are not
guaranteed to extract causally-relevant insights. In this work, we build on
ideas from the invariant causal prediction literature to propose Invariant
Causal Set Covering Machines, an extension of the classical Set Covering
Machine algorithm for conjunctions/disjunctions of binary-valued rules that
provably avoids spurious associations. We demonstrate both theoretically and
empirically that our method can identify the causal parents of a variable of
interest in polynomial time.
Related papers
- CausalConceptTS: Causal Attributions for Time Series Classification using High Fidelity Diffusion Models [1.068128849363198]
We introduce a novel framework to assess the causal effect of concepts on specific classification outcomes.
We leverage state-of-the-art diffusion-based generative models to estimate counterfactual outcomes.
Our approach compares these causal attributions with closely related associational attributions, both theoretically and empirically.
arXiv Detail & Related papers (2024-05-24T18:33:18Z) - Algorithmic syntactic causal identification [0.8901073744693314]
Causal identification in causal Bayes nets (CBNs) is an important tool in causal inference.
Most existing formulations of causal identification using techniques such as d-separation and do-calculus are expressed within the mathematical language of classical probability theory.
We show that this restriction can be lifted by replacing the use of classical probability theory with the alternative axiomatic foundation of symmetric monoidal categories.
arXiv Detail & Related papers (2024-03-14T17:14:53Z) - Nonparametric Partial Disentanglement via Mechanism Sparsity: Sparse
Actions, Interventions and Sparse Temporal Dependencies [58.179981892921056]
This work introduces a novel principle for disentanglement we call mechanism sparsity regularization.
We propose a representation learning method that induces disentanglement by simultaneously learning the latent factors.
We show that the latent factors can be recovered by regularizing the learned causal graph to be sparse.
arXiv Detail & Related papers (2024-01-10T02:38:21Z) - Algebraic Models for Qualified Aggregation in General Rough Sets, and
Reasoning Bias Discovery [0.0]
The research is motivated by the desire to model skeptical or pessimistic, and optimistic or possibilistic aggregation in human reasoning.
The model is suitable for the study of discriminatory/toxic behavior in human reasoning, and of ML algorithms learning such behavior.
arXiv Detail & Related papers (2023-08-30T17:07:54Z) - On Regularization and Inference with Label Constraints [62.60903248392479]
We compare two strategies for encoding label constraints in a machine learning pipeline, regularization with constraints and constrained inference.
For regularization, we show that it narrows the generalization gap by precluding models that are inconsistent with the constraints.
For constrained inference, we show that it reduces the population risk by correcting a model's violation, and hence turns the violation into an advantage.
arXiv Detail & Related papers (2023-07-08T03:39:22Z) - Exposing Attention Glitches with Flip-Flop Language Modeling [55.0688535574859]
This work identifies and analyzes the phenomenon of attention glitches in large language models.
We introduce flip-flop language modeling (FFLM), a family of synthetic benchmarks designed to probe the extrapolative behavior of neural language models.
We find that Transformer FFLMs suffer from a long tail of sporadic reasoning errors, some of which we can eliminate using various regularization techniques.
arXiv Detail & Related papers (2023-06-01T17:44:35Z) - Risk Measures and Upper Probabilities: Coherence and Stratification [7.88657961743755]
We look at richer alternatives to classical probability theory as a mathematical foundation for machine learning.
We examine a powerful and rich class of alternative aggregation functionals, known variously as spectral risk measures, Choquet integrals or Lorentz norms.
We empirically demonstrate how this new approach to uncertainty helps tackling practical machine learning problems.
arXiv Detail & Related papers (2022-06-07T11:08:16Z) - The Causal Marginal Polytope for Bounding Treatment Effects [9.196779204457059]
We propose a novel way to identify causal effects without constructing a global causal model.
We enforce compatibility between marginals of a causal model and data, without constructing a global causal model.
We call this collection of locally consistent marginals the causal marginal polytope.
arXiv Detail & Related papers (2022-02-28T15:08:22Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Consistency of mechanistic causal discovery in continuous-time using
Neural ODEs [85.7910042199734]
We consider causal discovery in continuous-time for the study of dynamical systems.
We propose a causal discovery algorithm based on penalized Neural ODEs.
arXiv Detail & Related papers (2021-05-06T08:48:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.