Safe Tests and Always-Valid Confidence Intervals for contingency tables
and beyond
- URL: http://arxiv.org/abs/2106.02693v1
- Date: Fri, 4 Jun 2021 20:12:13 GMT
- Title: Safe Tests and Always-Valid Confidence Intervals for contingency tables
and beyond
- Authors: Rosanne Turner, Alexander Ly, Peter Gr\"unwald
- Abstract summary: We develop E variables for testing whether two data streams come from the same source or not.
These E variables lead to tests that remain safe, under flexible sampling scenarios such as optional stopping and continuation.
- Score: 69.25055322530058
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We develop E variables for testing whether two data streams come from the
same source or not, and more generally, whether the difference between the
sources is larger than some minimal effect size. These E variables lead to
tests that remain safe, i.e. keep their Type-I error guarantees, under flexible
sampling scenarios such as optional stopping and continuation. We also develop
the corresponding always-valid confidence intervals. In special cases our E
variables also have an optimal `growth' property under the alternative. We
illustrate the generic construction through the special case of 2x2 contingency
tables, where we also allow for the incorporation of different restrictions on
a composite alternative. Comparison to p-value analysis in simulations and a
real-world example show that E variables, through their flexibility, often
allow for early stopping of data collection, thereby retaining similar power as
classical methods.
Related papers
- Diversity-aware Buffer for Coping with Temporally Correlated Data
Streams in Online Test-time Adaptation [3.1265626879839923]
Test data streams are not always independent and identically distributed (i.i.d.)
We propose a diversity-aware and category-balanced buffer that can simulate an i.i.d. data stream, even in non-i.i.d. scenarios.
We achieve state-of-the-art results on most considered benchmarks.
arXiv Detail & Related papers (2024-01-02T01:56:25Z) - Non-Exchangeable Conformal Risk Control [12.381447108228635]
Split conformal prediction has recently sparked great interest due to its ability to provide formally guaranteed uncertainty sets or intervals.
We propose non-exchangeable conformal risk control, which allows controlling the expected value of any monotone loss function when the data is not exchangeable.
Our framework is flexible, makes very few assumptions, and allows weighting the data based on its relevance for a given test example.
arXiv Detail & Related papers (2023-10-02T15:00:19Z) - Robust Continual Test-time Adaptation: Instance-aware BN and
Prediction-balanced Memory [58.72445309519892]
We present a new test-time adaptation scheme that is robust against non-i.i.d. test data streams.
Our novelty is mainly two-fold: (a) Instance-Aware Batch Normalization (IABN) that corrects normalization for out-of-distribution samples, and (b) Prediction-balanced Reservoir Sampling (PBRS) that simulates i.i.d. data stream from non-i.i.d. stream in a class-balanced manner.
arXiv Detail & Related papers (2022-08-10T03:05:46Z) - Predicting Out-of-Domain Generalization with Neighborhood Invariance [59.05399533508682]
We propose a measure of a classifier's output invariance in a local transformation neighborhood.
Our measure is simple to calculate, does not depend on the test point's true label, and can be applied even in out-of-domain (OOD) settings.
In experiments on benchmarks in image classification, sentiment analysis, and natural language inference, we demonstrate a strong and robust correlation between our measure and actual OOD generalization.
arXiv Detail & Related papers (2022-07-05T14:55:16Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - Multivariate Probabilistic Regression with Natural Gradient Boosting [63.58097881421937]
We propose a Natural Gradient Boosting (NGBoost) approach based on nonparametrically modeling the conditional parameters of the multivariate predictive distribution.
Our method is robust, works out-of-the-box without extensive tuning, is modular with respect to the assumed target distribution, and performs competitively in comparison to existing approaches.
arXiv Detail & Related papers (2021-06-07T17:44:49Z) - Stable Prediction via Leveraging Seed Variable [73.9770220107874]
Previous machine learning methods might exploit subtly spurious correlations in training data induced by non-causal variables for prediction.
We propose a conditional independence test based algorithm to separate causal variables with a seed variable as priori, and adopt them for stable prediction.
Our algorithm outperforms state-of-the-art methods for stable prediction.
arXiv Detail & Related papers (2020-06-09T06:56:31Z) - Safe Testing [0.9634859579172255]
We develop the theory of hypothesis testing based on the e-value.
Tests based on e-values are safe, i.e. they preserve Type-I error guarantees.
arXiv Detail & Related papers (2019-06-18T20:39:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.