Empirical Frequentist Coverage of Deep Learning Uncertainty
Quantification Procedures
- URL: http://arxiv.org/abs/2010.03039v2
- Date: Wed, 24 Feb 2021 20:54:18 GMT
- Title: Empirical Frequentist Coverage of Deep Learning Uncertainty
Quantification Procedures
- Authors: Benjamin Kompa, Jasper Snoek, Andrew Beam
- Abstract summary: We provide the first large scale evaluation of the empirical frequentist coverage properties of uncertainty quantification techniques.
We find that, in general, some methods do achieve desirable coverage properties on in distribution samples, but that coverage is not maintained on out-of-distribution data.
- Score: 13.890139530120164
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Uncertainty quantification for complex deep learning models is increasingly
important as these techniques see growing use in high-stakes, real-world
settings. Currently, the quality of a model's uncertainty is evaluated using
point-prediction metrics such as negative log-likelihood or the Brier score on
heldout data. In this study, we provide the first large scale evaluation of the
empirical frequentist coverage properties of well known uncertainty
quantification techniques on a suite of regression and classification tasks. We
find that, in general, some methods do achieve desirable coverage properties on
in distribution samples, but that coverage is not maintained on
out-of-distribution data. Our results demonstrate the failings of current
uncertainty quantification techniques as dataset shift increases and establish
coverage as an important metric in developing models for real-world
applications.
Related papers
- Non-Asymptotic Uncertainty Quantification in High-Dimensional Learning [5.318766629972959]
Uncertainty quantification is a crucial but challenging task in many high-dimensional regression or learning problems.
We develop a new data-driven approach for UQ in regression that applies both to classical regression approaches as well as to neural networks.
arXiv Detail & Related papers (2024-07-18T16:42:10Z) - One step closer to unbiased aleatoric uncertainty estimation [71.55174353766289]
We propose a new estimation method by actively de-noising the observed data.
By conducting a broad range of experiments, we demonstrate that our proposed approach provides a much closer approximation to the actual data uncertainty than the standard method.
arXiv Detail & Related papers (2023-12-16T14:59:11Z) - Estimating Uncertainty in Landslide Segmentation Models [7.537865319452023]
Landslides are a recurring, widespread hazard. Preparation and mitigation efforts can be aided by a high-quality, large-scale dataset that covers global at-risk areas.
Recent automated efforts focus on deep learning models for landslide segmentation from satellite imagery.
Accurate and robust uncertainty estimates can enable low-cost oversight of auto-generated landslide databases to resolve errors, identify hard negative examples, and increase the size of labeled training data.
arXiv Detail & Related papers (2023-11-18T18:18:33Z) - Uncertainty Estimation by Fisher Information-based Evidential Deep
Learning [61.94125052118442]
Uncertainty estimation is a key factor that makes deep learning reliable in practical applications.
We propose a novel method, Fisher Information-based Evidential Deep Learning ($mathcalI$-EDL)
In particular, we introduce Fisher Information Matrix (FIM) to measure the informativeness of evidence carried by each sample, according to which we can dynamically reweight the objective loss terms to make the network more focused on the representation learning of uncertain classes.
arXiv Detail & Related papers (2023-03-03T16:12:59Z) - Benchmarking common uncertainty estimation methods with
histopathological images under domain shift and label noise [62.997667081978825]
In high-risk environments, deep learning models need to be able to judge their uncertainty and reject inputs when there is a significant chance of misclassification.
We conduct a rigorous evaluation of the most commonly used uncertainty and robustness methods for the classification of Whole Slide Images.
We observe that ensembles of methods generally lead to better uncertainty estimates as well as an increased robustness towards domain shifts and label noise.
arXiv Detail & Related papers (2023-01-03T11:34:36Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Know Your Limits: Uncertainty Estimation with ReLU Classifiers Fails at
Reliable OOD Detection [0.0]
This paper gives a theoretical explanation for said experimental findings and illustrates it on synthetic data.
We prove that such techniques are not able to reliably identify OOD samples in a classification setting.
arXiv Detail & Related papers (2020-12-09T21:35:55Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.