Probabilities of Causation: Adequate Size of Experimental and
Observational Samples
- URL: http://arxiv.org/abs/2210.05027v1
- Date: Mon, 10 Oct 2022 21:59:49 GMT
- Title: Probabilities of Causation: Adequate Size of Experimental and
Observational Samples
- Authors: Ang Li, Ruirui Mao, Judea Pearl
- Abstract summary: Tian and Pearl derived sharp bounds for the probability of necessity and sufficiency (PNS), the probability of sufficiency (PS), and the probability of necessity (PN) using experimental and observational data.
The assumption is that one is in possession of a large enough sample to permit an accurate estimation of the experimental and observational distributions.
We present a method for determining the sample size needed for such estimation, when a given confidence interval (CI) is specified.
- Score: 17.565045120151865
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The probabilities of causation are commonly used to solve decision-making
problems. Tian and Pearl derived sharp bounds for the probability of necessity
and sufficiency (PNS), the probability of sufficiency (PS), and the probability
of necessity (PN) using experimental and observational data. The assumption is
that one is in possession of a large enough sample to permit an accurate
estimation of the experimental and observational distributions. In this study,
we present a method for determining the sample size needed for such estimation,
when a given confidence interval (CI) is specified. We further show by
simulation that the proposed sample size delivered stable estimations of the
bounds of PNS.
Related papers
- To Believe or Not to Believe Your LLM [51.2579827761899]
We explore uncertainty quantification in large language models (LLMs)
We derive an information-theoretic metric that allows to reliably detect when only epistemic uncertainty is large.
We conduct a series of experiments which demonstrate the advantage of our formulation.
arXiv Detail & Related papers (2024-06-04T17:58:18Z) - Intervention and Conditioning in Causal Bayesian Networks [23.225006087292765]
We show that by making simple yet often realistic independence assumptions, it is possible to estimate the probability of an interventional formula.
In many cases of interest, when the assumptions are appropriate, these probability estimates can be evaluated using observational data.
arXiv Detail & Related papers (2024-05-23T15:55:38Z) - User-defined Event Sampling and Uncertainty Quantification in Diffusion
Models for Physical Dynamical Systems [49.75149094527068]
We show that diffusion models can be adapted to make predictions and provide uncertainty quantification for chaotic dynamical systems.
We develop a probabilistic approximation scheme for the conditional score function which converges to the true distribution as the noise level decreases.
We are able to sample conditionally on nonlinear userdefined events at inference time, and matches data statistics even when sampling from the tails of the distribution.
arXiv Detail & Related papers (2023-06-13T03:42:03Z) - Detecting Adversarial Data by Probing Multiple Perturbations Using
Expected Perturbation Score [62.54911162109439]
Adversarial detection aims to determine whether a given sample is an adversarial one based on the discrepancy between natural and adversarial distributions.
We propose a new statistic called expected perturbation score (EPS), which is essentially the expected score of a sample after various perturbations.
We develop EPS-based maximum mean discrepancy (MMD) as a metric to measure the discrepancy between the test sample and natural samples.
arXiv Detail & Related papers (2023-05-25T13:14:58Z) - Probabilities of Causation: Role of Observational Data [20.750773939911685]
We discuss the conditions that observational data are worth considering to improve the quality of the bounds.
We also apply the proposed theorems to the unit selection problem defined by Li and Pearl.
arXiv Detail & Related papers (2022-10-17T09:10:11Z) - Counterfactual Inference of the Mean Outcome under a Convergence of
Average Logging Probability [5.596752018167751]
This paper considers estimating the mean outcome of an action from samples obtained in adaptive experiments.
In adaptive experiments, the probability of choosing an action is allowed to be sequentially updated based on past observations.
arXiv Detail & Related papers (2021-02-17T19:05:53Z) - Tracking disease outbreaks from sparse data with Bayesian inference [55.82986443159948]
The COVID-19 pandemic provides new motivation for estimating the empirical rate of transmission during an outbreak.
Standard methods struggle to accommodate the partial observability and sparse data common at finer scales.
We propose a Bayesian framework which accommodates partial observability in a principled manner.
arXiv Detail & Related papers (2020-09-12T20:37:33Z) - Density of States Estimation for Out-of-Distribution Detection [69.90130863160384]
DoSE is the density of states estimator.
We demonstrate DoSE's state-of-the-art performance against other unsupervised OOD detectors.
arXiv Detail & Related papers (2020-06-16T16:06:25Z) - Balance-Subsampled Stable Prediction [55.13512328954456]
We propose a novel balance-subsampled stable prediction (BSSP) algorithm based on the theory of fractional factorial design.
A design-theoretic analysis shows that the proposed method can reduce the confounding effects among predictors induced by the distribution shift.
Numerical experiments on both synthetic and real-world data sets demonstrate that our BSSP algorithm significantly outperforms the baseline methods for stable prediction across unknown test data.
arXiv Detail & Related papers (2020-06-08T07:01:38Z) - Witnessing Entanglement in Experiments with Correlated Noise [1.1246250197597698]
We propose two methods to analyze witness experiments where the states can be subject to arbitrarily correlated noise.
The first method is a rejection experiment, in which we certify the creation of entanglement by rejecting the hypothesis that the experiment can only produce separable states.
The second method is an estimation experiment, in which we estimate and construct confidence intervals for the average witness value.
arXiv Detail & Related papers (2020-02-27T19:23:08Z) - Estimating the number and effect sizes of non-null hypotheses [14.34147140416535]
Knowing the distribution of effect sizes allows us to calculate the power (type II error) of any experimental design.
Our estimator can be used to guarantee the number of discoveries that will be made using a given experimental design in a future experiment.
arXiv Detail & Related papers (2020-02-17T23:20:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.