Uncertainty Estimates of Predictions via a General Bias-Variance
Decomposition
- URL: http://arxiv.org/abs/2210.12256v3
- Date: Thu, 20 Apr 2023 21:01:19 GMT
- Title: Uncertainty Estimates of Predictions via a General Bias-Variance
Decomposition
- Authors: Sebastian G. Gruber, Florian Buettner
- Abstract summary: We introduce a bias-variance decomposition for proper scores, giving rise to the Bregman Information as the variance term.
We showcase the practical relevance of this decomposition on several downstream tasks, including model ensembles and confidence regions.
- Score: 7.811916700683125
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reliably estimating the uncertainty of a prediction throughout the model
lifecycle is crucial in many safety-critical applications. The most common way
to measure this uncertainty is via the predicted confidence. While this tends
to work well for in-domain samples, these estimates are unreliable under domain
drift and restricted to classification. Alternatively, proper scores can be
used for most predictive tasks but a bias-variance decomposition for model
uncertainty does not exist in the current literature. In this work we introduce
a general bias-variance decomposition for proper scores, giving rise to the
Bregman Information as the variance term. We discover how exponential families
and the classification log-likelihood are special cases and provide novel
formulations. Surprisingly, we can express the classification case purely in
the logit space. We showcase the practical relevance of this decomposition on
several downstream tasks, including model ensembles and confidence regions.
Further, we demonstrate how different approximations of the instance-level
Bregman Information allow reliable out-of-distribution detection for all
degrees of domain drift.
Related papers
- Evaluating AI systems under uncertain ground truth: a case study in
dermatology [44.80772162289557]
We propose a metric for measuring annotation uncertainty and provide uncertainty-adjusted metrics for performance evaluation.
We present a case study applying our framework to skin condition classification from images where annotations are provided in the form of differential diagnoses.
arXiv Detail & Related papers (2023-07-05T10:33:45Z) - How Reliable is Your Regression Model's Uncertainty Under Real-World
Distribution Shifts? [46.05502630457458]
We propose a benchmark of 8 image-based regression datasets with different types of challenging distribution shifts.
We find that while methods are well calibrated when there is no distribution shift, they all become highly overconfident on many of the benchmark datasets.
arXiv Detail & Related papers (2023-02-07T18:54:39Z) - Confidence and Dispersity Speak: Characterising Prediction Matrix for
Unsupervised Accuracy Estimation [51.809741427975105]
This work aims to assess how well a model performs under distribution shifts without using labels.
We use the nuclear norm that has been shown to be effective in characterizing both properties.
We show that the nuclear norm is more accurate and robust in accuracy than existing methods.
arXiv Detail & Related papers (2023-02-02T13:30:48Z) - A Statistical Model for Predicting Generalization in Few-Shot
Classification [6.158812834002346]
We introduce a Gaussian model of the feature distribution to predict the generalization error.
We show that our approach outperforms alternatives such as the leave-one-out cross-validation strategy.
arXiv Detail & Related papers (2022-12-13T10:21:15Z) - The Implicit Delta Method [61.36121543728134]
In this paper, we propose an alternative, the implicit delta method, which works by infinitesimally regularizing the training loss of uncertainty.
We show that the change in the evaluation due to regularization is consistent for the variance of the evaluation estimator, even when the infinitesimal change is approximated by a finite difference.
arXiv Detail & Related papers (2022-11-11T19:34:17Z) - Uncertainty Estimation for Multi-view Data: The Power of Seeing the
Whole Picture [5.868139834982011]
Uncertainty estimation is essential to make neural networks trustworthy in real-world applications.
We propose a new multi-view classification framework for better uncertainty estimation and out-of-domain sample detection.
arXiv Detail & Related papers (2022-10-06T04:47:51Z) - Uncertainty Modeling for Out-of-Distribution Generalization [56.957731893992495]
We argue that the feature statistics can be properly manipulated to improve the generalization ability of deep learning models.
Common methods often consider the feature statistics as deterministic values measured from the learned features.
We improve the network generalization ability by modeling the uncertainty of domain shifts with synthesized feature statistics during training.
arXiv Detail & Related papers (2022-02-08T16:09:12Z) - Dense Uncertainty Estimation via an Ensemble-based Conditional Latent
Variable Model [68.34559610536614]
We argue that the aleatoric uncertainty is an inherent attribute of the data and can only be correctly estimated with an unbiased oracle model.
We propose a new sampling and selection strategy at train time to approximate the oracle model for aleatoric uncertainty estimation.
Our results show that our solution achieves both accurate deterministic results and reliable uncertainty estimation.
arXiv Detail & Related papers (2021-11-22T08:54:10Z) - Learning Probabilistic Ordinal Embeddings for Uncertainty-Aware
Regression [91.3373131262391]
Uncertainty is the only certainty there is.
Traditionally, the direct regression formulation is considered and the uncertainty is modeled by modifying the output space to a certain family of probabilistic distributions.
How to model the uncertainty within the present-day technologies for regression remains an open issue.
arXiv Detail & Related papers (2021-03-25T06:56:09Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.