Test-time Recalibration of Conformal Predictors Under Distribution Shift
Based on Unlabeled Examples
- URL: http://arxiv.org/abs/2210.04166v2
- Date: Sat, 3 Jun 2023 21:34:49 GMT
- Title: Test-time Recalibration of Conformal Predictors Under Distribution Shift
Based on Unlabeled Examples
- Authors: Fatih Furkan Yilmaz and Reinhard Heckel
- Abstract summary: Conformal predictors provide uncertainty estimates by computing a set of classes with a user-specified probability.
We propose a method that provides excellent uncertainty estimates under natural distribution shifts.
- Score: 30.61588337557343
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern image classifiers are very accurate, but the predictions come without
uncertainty estimates. Conformal predictors provide uncertainty estimates by
computing a set of classes containing the correct class with a user-specified
probability based on the classifier's probability estimates. To provide such
sets, conformal predictors often estimate a cutoff threshold for the
probability estimates based on a calibration set. Conformal predictors
guarantee reliability only when the calibration set is from the same
distribution as the test set. Therefore, conformal predictors need to be
recalibrated for new distributions. However, in practice, labeled data from new
distributions is rarely available, making calibration infeasible. In this work,
we consider the problem of predicting the cutoff threshold for a new
distribution based on unlabeled examples. While it is impossible in general to
guarantee reliability when calibrating based on unlabeled examples, we propose
a method that provides excellent uncertainty estimates under natural
distribution shifts, and provably works for a specific model of a distribution
shift.
Related papers
- Provably Reliable Conformal Prediction Sets in the Presence of Data Poisoning [53.42244686183879]
Conformal prediction provides model-agnostic and distribution-free uncertainty quantification.
Yet, conformal prediction is not reliable under poisoning attacks where adversaries manipulate both training and calibration data.
We propose reliable prediction sets (RPS): the first efficient method for constructing conformal prediction sets with provable reliability guarantees under poisoning.
arXiv Detail & Related papers (2024-10-13T15:37:11Z) - Quantification of Predictive Uncertainty via Inference-Time Sampling [57.749601811982096]
We propose a post-hoc sampling strategy for estimating predictive uncertainty accounting for data ambiguity.
The method can generate different plausible outputs for a given input and does not assume parametric forms of predictive distributions.
arXiv Detail & Related papers (2023-08-03T12:43:21Z) - Calibrated Selective Classification [34.08454890436067]
We develop a new approach to selective classification in which we propose a method for rejecting examples with "uncertain" uncertainties.
We present a framework for learning selectively calibrated models, where a separate selector network is trained to improve the selective calibration error of a given base model.
We demonstrate the empirical effectiveness of our approach on multiple image classification and lung cancer risk assessment tasks.
arXiv Detail & Related papers (2022-08-25T13:31:09Z) - On the Calibration of Probabilistic Classifier Sets [6.759124697337311]
We extend the notion of calibration to evaluate the validity of an aleatoric uncertainty representation.
We show that ensembles of deep neural networks are often not well calibrated.
arXiv Detail & Related papers (2022-05-20T10:57:46Z) - Training on Test Data with Bayesian Adaptation for Covariate Shift [96.3250517412545]
Deep neural networks often make inaccurate predictions with unreliable uncertainty estimates.
We derive a Bayesian model that provides for a well-defined relationship between unlabeled inputs under distributional shift and model parameters.
We show that our method improves both accuracy and uncertainty estimation.
arXiv Detail & Related papers (2021-09-27T01:09:08Z) - Distribution-free uncertainty quantification for classification under
label shift [105.27463615756733]
We focus on uncertainty quantification (UQ) for classification problems via two avenues.
We first argue that label shift hurts UQ, by showing degradation in coverage and calibration.
We examine these techniques theoretically in a distribution-free framework and demonstrate their excellent practical performance.
arXiv Detail & Related papers (2021-03-04T20:51:03Z) - Distribution-free binary classification: prediction sets, confidence
intervals and calibration [106.50279469344937]
We study three notions of uncertainty quantification -- calibration, confidence intervals and prediction sets -- for binary classification in the distribution-free setting.
We derive confidence intervals for binned probabilities for both fixed-width and uniform-mass binning.
As a consequence of our 'tripod' theorems, these confidence intervals for binned probabilities lead to distribution-free calibration.
arXiv Detail & Related papers (2020-06-18T14:17:29Z) - Individual Calibration with Randomized Forecasting [116.2086707626651]
We show that calibration for individual samples is possible in the regression setup if the predictions are randomized.
We design a training objective to enforce individual calibration and use it to train randomized regression functions.
arXiv Detail & Related papers (2020-06-18T05:53:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.