Test-time Recalibration of Conformal Predictors Under Distribution Shift
Based on Unlabeled Examples
- URL: http://arxiv.org/abs/2210.04166v2
- Date: Sat, 3 Jun 2023 21:34:49 GMT
- Title: Test-time Recalibration of Conformal Predictors Under Distribution Shift
Based on Unlabeled Examples
- Authors: Fatih Furkan Yilmaz and Reinhard Heckel
- Abstract summary: Conformal predictors provide uncertainty estimates by computing a set of classes with a user-specified probability.
We propose a method that provides excellent uncertainty estimates under natural distribution shifts.
- Score: 30.61588337557343
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern image classifiers are very accurate, but the predictions come without
uncertainty estimates. Conformal predictors provide uncertainty estimates by
computing a set of classes containing the correct class with a user-specified
probability based on the classifier's probability estimates. To provide such
sets, conformal predictors often estimate a cutoff threshold for the
probability estimates based on a calibration set. Conformal predictors
guarantee reliability only when the calibration set is from the same
distribution as the test set. Therefore, conformal predictors need to be
recalibrated for new distributions. However, in practice, labeled data from new
distributions is rarely available, making calibration infeasible. In this work,
we consider the problem of predicting the cutoff threshold for a new
distribution based on unlabeled examples. While it is impossible in general to
guarantee reliability when calibrating based on unlabeled examples, we propose
a method that provides excellent uncertainty estimates under natural
distribution shifts, and provably works for a specific model of a distribution
shift.
Related papers
- Sparse Activations as Conformal Predictors [19.298282860984116]
We find a novel connection between conformal prediction and sparse softmax-like transformations.
We introduce new non-conformity scores for classification that make the calibration process correspond to the widely used temperature scaling method.
We show that the proposed method achieves competitive results in terms of coverage, efficiency, and adaptiveness.
arXiv Detail & Related papers (2025-02-20T17:53:41Z) - Conformal Prediction Sets with Improved Conditional Coverage using Trust Scores [52.92618442300405]
It is impossible to achieve exact, distribution-free conditional coverage in finite samples.
We propose an alternative conformal prediction algorithm that targets coverage where it matters most.
arXiv Detail & Related papers (2025-01-17T12:01:56Z) - Provable Uncertainty Decomposition via Higher-Order Calibration [4.969075533165688]
We give a principled method for decomposing the predictive uncertainty of a model into aleatoric and epistemic components.
Our method is based on the new notion of higher-order calibration.
We demonstrate through experiments that our method produces meaningful uncertainty decompositions for image classification.
arXiv Detail & Related papers (2024-12-25T07:26:36Z) - Provably Reliable Conformal Prediction Sets in the Presence of Data Poisoning [53.42244686183879]
Conformal prediction provides model-agnostic and distribution-free uncertainty quantification.
Yet, conformal prediction is not reliable under poisoning attacks where adversaries manipulate both training and calibration data.
We propose reliable prediction sets (RPS): the first efficient method for constructing conformal prediction sets with provable reliability guarantees under poisoning.
arXiv Detail & Related papers (2024-10-13T15:37:11Z) - Calibrated Selective Classification [34.08454890436067]
We develop a new approach to selective classification in which we propose a method for rejecting examples with "uncertain" uncertainties.
We present a framework for learning selectively calibrated models, where a separate selector network is trained to improve the selective calibration error of a given base model.
We demonstrate the empirical effectiveness of our approach on multiple image classification and lung cancer risk assessment tasks.
arXiv Detail & Related papers (2022-08-25T13:31:09Z) - Training on Test Data with Bayesian Adaptation for Covariate Shift [96.3250517412545]
Deep neural networks often make inaccurate predictions with unreliable uncertainty estimates.
We derive a Bayesian model that provides for a well-defined relationship between unlabeled inputs under distributional shift and model parameters.
We show that our method improves both accuracy and uncertainty estimation.
arXiv Detail & Related papers (2021-09-27T01:09:08Z) - Distribution-free uncertainty quantification for classification under
label shift [105.27463615756733]
We focus on uncertainty quantification (UQ) for classification problems via two avenues.
We first argue that label shift hurts UQ, by showing degradation in coverage and calibration.
We examine these techniques theoretically in a distribution-free framework and demonstrate their excellent practical performance.
arXiv Detail & Related papers (2021-03-04T20:51:03Z) - Distribution-free binary classification: prediction sets, confidence
intervals and calibration [106.50279469344937]
We study three notions of uncertainty quantification -- calibration, confidence intervals and prediction sets -- for binary classification in the distribution-free setting.
We derive confidence intervals for binned probabilities for both fixed-width and uniform-mass binning.
As a consequence of our 'tripod' theorems, these confidence intervals for binned probabilities lead to distribution-free calibration.
arXiv Detail & Related papers (2020-06-18T14:17:29Z) - Individual Calibration with Randomized Forecasting [116.2086707626651]
We show that calibration for individual samples is possible in the regression setup if the predictions are randomized.
We design a training objective to enforce individual calibration and use it to train randomized regression functions.
arXiv Detail & Related papers (2020-06-18T05:53:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.