Confidence Sets and Hypothesis Testing in a Likelihood-Free Inference
Setting
- URL: http://arxiv.org/abs/2002.10399v2
- Date: Fri, 14 Aug 2020 02:56:38 GMT
- Title: Confidence Sets and Hypothesis Testing in a Likelihood-Free Inference
Setting
- Authors: Niccol\`o Dalmasso and Rafael Izbicki and Ann B. Lee
- Abstract summary: $texttACORE$ is a frequentist approach to LFI that first formulates the classical likelihood ratio test (LRT) as a parametrized classification problem.
$texttACORE$ is based on the key observation that the statistic, the rejection probability of the test, and the coverage of the confidence set are conditional distribution functions.
- Score: 5.145741425164947
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Parameter estimation, statistical tests and confidence sets are the
cornerstones of classical statistics that allow scientists to make inferences
about the underlying process that generated the observed data. A key question
is whether one can still construct hypothesis tests and confidence sets with
proper coverage and high power in a so-called likelihood-free inference (LFI)
setting; that is, a setting where the likelihood is not explicitly known but
one can forward-simulate observable data according to a stochastic model. In
this paper, we present $\texttt{ACORE}$ (Approximate Computation via Odds Ratio
Estimation), a frequentist approach to LFI that first formulates the classical
likelihood ratio test (LRT) as a parametrized classification problem, and then
uses the equivalence of tests and confidence sets to build confidence regions
for parameters of interest. We also present a goodness-of-fit procedure for
checking whether the constructed tests and confidence regions are valid.
$\texttt{ACORE}$ is based on the key observation that the LRT statistic, the
rejection probability of the test, and the coverage of the confidence set are
conditional distribution functions which often vary smoothly as a function of
the parameters of interest. Hence, instead of relying solely on samples
simulated at fixed parameter settings (as is the convention in standard Monte
Carlo solutions), one can leverage machine learning tools and data simulated in
the neighborhood of a parameter to improve estimates of quantities of interest.
We demonstrate the efficacy of $\texttt{ACORE}$ with both theoretical and
empirical results. Our implementation is available on Github.
Related papers
- Extended Fiducial Inference: Toward an Automated Process of Statistical Inference [9.277340234795801]
We develop a new statistical inference method called extended Fiducial inference (EFI)
The new method achieves the goal of fiducial inference by leveraging advanced statistical computing techniques.
EFI offers significant advantages in parameter estimation and hypothesis testing.
arXiv Detail & Related papers (2024-07-31T14:15:42Z) - Mitigating LLM Hallucinations via Conformal Abstention [70.83870602967625]
We develop a principled procedure for determining when a large language model should abstain from responding in a general domain.
We leverage conformal prediction techniques to develop an abstention procedure that benefits from rigorous theoretical guarantees on the hallucination rate (error rate)
Experimentally, our resulting conformal abstention method reliably bounds the hallucination rate on various closed-book, open-domain generative question answering datasets.
arXiv Detail & Related papers (2024-04-04T11:32:03Z) - Finite Sample Confidence Regions for Linear Regression Parameters Using
Arbitrary Predictors [1.6860963320038902]
We explore a novel methodology for constructing confidence regions for parameters of linear models, using predictions from any arbitrary predictor.
The derived confidence regions can be cast as constraints within a Mixed Linear Programming framework, enabling optimisation of linear objectives.
Unlike previous methods, the confidence region can be empty, which can be used for hypothesis testing.
arXiv Detail & Related papers (2024-01-27T00:15:48Z) - SMURF-THP: Score Matching-based UnceRtainty quantiFication for
Transformer Hawkes Process [76.98721879039559]
We propose SMURF-THP, a score-based method for learning Transformer Hawkes process and quantifying prediction uncertainty.
Specifically, SMURF-THP learns the score function of events' arrival time based on a score-matching objective.
We conduct extensive experiments in both event type prediction and uncertainty quantification of arrival time.
arXiv Detail & Related papers (2023-10-25T03:33:45Z) - Calibrating Neural Simulation-Based Inference with Differentiable
Coverage Probability [50.44439018155837]
We propose to include a calibration term directly into the training objective of the neural model.
By introducing a relaxation of the classical formulation of calibration error we enable end-to-end backpropagation.
It is directly applicable to existing computational pipelines allowing reliable black-box posterior inference.
arXiv Detail & Related papers (2023-10-20T10:20:45Z) - When Does Confidence-Based Cascade Deferral Suffice? [69.28314307469381]
Cascades are a classical strategy to enable inference cost to vary adaptively across samples.
A deferral rule determines whether to invoke the next classifier in the sequence, or to terminate prediction.
Despite being oblivious to the structure of the cascade, confidence-based deferral often works remarkably well in practice.
arXiv Detail & Related papers (2023-07-06T04:13:57Z) - Stable Probability Weighting: Large-Sample and Finite-Sample Estimation
and Inference Methods for Heterogeneous Causal Effects of Multivalued
Treatments Under Limited Overlap [0.0]
I propose new practical large-sample and finite-sample methods for estimating and inferring heterogeneous causal effects.
I develop a general principle called "Stable Probability Weighting"
I also propose new finite-sample inference methods for testing a general class of weak null hypotheses.
arXiv Detail & Related papers (2023-01-13T18:52:18Z) - Nonparametric Conditional Local Independence Testing [69.31200003384122]
Conditional local independence is an independence relation among continuous time processes.
No nonparametric test of conditional local independence has been available.
We propose such a nonparametric test based on double machine learning.
arXiv Detail & Related papers (2022-03-25T10:31:02Z) - Likelihood-Free Frequentist Inference: Bridging Classical Statistics and Machine Learning for Reliable Simulator-Based Inference [4.525512100042707]
We propose a modular inference framework that bridges classical statistics and modern machine learning.
We refer to this framework as likelihood-free frequentist inference (LF2I)
arXiv Detail & Related papers (2021-07-08T15:52:18Z) - Testing for Outliers with Conformal p-values [14.158078752410182]
The goal is to test whether new independent samples belong to the same distribution as a reference data set or are outliers.
We propose a solution based on conformal inference, a broadly applicable framework which yields p-values that are marginally valid but mutually dependent for different test points.
We prove these p-values are positively dependent and enable exact false discovery rate control, although in a relatively weak marginal sense.
arXiv Detail & Related papers (2021-04-16T17:59:21Z) - Binary Classification from Positive Data with Skewed Confidence [85.18941440826309]
Positive-confidence (Pconf) classification is a promising weakly-supervised learning method.
In practice, the confidence may be skewed by bias arising in an annotation process.
We introduce the parameterized model of the skewed confidence, and propose the method for selecting the hyper parameter.
arXiv Detail & Related papers (2020-01-29T00:04:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.