Estimation of Accurate and Calibrated Uncertainties in Deterministic
models
- URL: http://arxiv.org/abs/2003.05103v1
- Date: Wed, 11 Mar 2020 04:02:56 GMT
- Title: Estimation of Accurate and Calibrated Uncertainties in Deterministic
models
- Authors: Enrico Camporeale and Algo Car\`e
- Abstract summary: We devise a method to transform a deterministic prediction into a probabilistic one.
We show that for doing so, one has to compromise between the accuracy and the reliability (calibration) of such a model.
We show several examples both with synthetic data, where the underlying hidden noise can accurately be recovered, and with large real-world datasets.
- Score: 0.8702432681310401
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper we focus on the problem of assigning uncertainties to
single-point predictions generated by a deterministic model that outputs a
continuous variable. This problem applies to any state-of-the-art physics or
engineering models that have a computational cost that does not readily allow
to run ensembles and to estimate the uncertainty associated to single-point
predictions. Essentially, we devise a method to easily transform a
deterministic prediction into a probabilistic one. We show that for doing so,
one has to compromise between the accuracy and the reliability (calibration) of
such a probabilistic model. Hence, we introduce a cost function that encodes
their trade-off. We use the Continuous Rank Probability Score to measure
accuracy and we derive an analytic formula for the reliability, in the case of
forecasts of continuous scalar variables expressed in terms of Gaussian
distributions. The new Accuracy-Reliability cost function is then used to
estimate the input-dependent variance, given a black-box mean function, by
solving a two-objective optimization problem. The simple philosophy behind this
strategy is that predictions based on the estimated variances should not only
be accurate, but also reliable (i.e. statistical consistent with observations).
Conversely, early works based on the minimization of classical cost functions,
such as the negative log probability density, cannot simultaneously enforce
both accuracy and reliability. We show several examples both with synthetic
data, where the underlying hidden noise can accurately be recovered, and with
large real-world datasets.
Related papers
- Invariant Probabilistic Prediction [45.90606906307022]
We show that arbitrary distribution shifts do not, in general, admit invariant and robust probabilistic predictions.
We propose a method to yield invariant probabilistic predictions, called IPP, and study the consistency of the underlying parameters.
arXiv Detail & Related papers (2023-09-18T18:50:24Z) - Sparsified Simultaneous Confidence Intervals for High-Dimensional Linear
Models [4.010566541114989]
We propose a notion of simultaneous confidence intervals called the sparsified simultaneous confidence intervals.
Our intervals are sparse in the sense that some of the intervals' upper and lower bounds are shrunken to zero.
The proposed method can be coupled with various selection procedures, making it ideal for comparing their uncertainty.
arXiv Detail & Related papers (2023-07-14T18:37:57Z) - The Implicit Delta Method [61.36121543728134]
In this paper, we propose an alternative, the implicit delta method, which works by infinitesimally regularizing the training loss of uncertainty.
We show that the change in the evaluation due to regularization is consistent for the variance of the evaluation estimator, even when the infinitesimal change is approximated by a finite difference.
arXiv Detail & Related papers (2022-11-11T19:34:17Z) - Metrics of calibration for probabilistic predictions [0.0]
"Reliability diagrams" help detect and diagnose statistically significant discrepancies -- so-called "miscalibration"
The canonical reliability diagrams histogram the observed and expected values of the predictions.
But, which widths of bins or kernels are best?
Slope is easy to perceive with quantitative precision, even when the constant offsets of the secant lines are irrelevant.
arXiv Detail & Related papers (2022-05-19T16:38:24Z) - Dense Uncertainty Estimation via an Ensemble-based Conditional Latent
Variable Model [68.34559610536614]
We argue that the aleatoric uncertainty is an inherent attribute of the data and can only be correctly estimated with an unbiased oracle model.
We propose a new sampling and selection strategy at train time to approximate the oracle model for aleatoric uncertainty estimation.
Our results show that our solution achieves both accurate deterministic results and reliable uncertainty estimation.
arXiv Detail & Related papers (2021-11-22T08:54:10Z) - Dense Uncertainty Estimation [62.23555922631451]
In this paper, we investigate neural networks and uncertainty estimation techniques to achieve both accurate deterministic prediction and reliable uncertainty estimation.
We work on two types of uncertainty estimations solutions, namely ensemble based methods and generative model based methods, and explain their pros and cons while using them in fully/semi/weakly-supervised framework.
arXiv Detail & Related papers (2021-10-13T01:23:48Z) - Multivariate Probabilistic Regression with Natural Gradient Boosting [63.58097881421937]
We propose a Natural Gradient Boosting (NGBoost) approach based on nonparametrically modeling the conditional parameters of the multivariate predictive distribution.
Our method is robust, works out-of-the-box without extensive tuning, is modular with respect to the assumed target distribution, and performs competitively in comparison to existing approaches.
arXiv Detail & Related papers (2021-06-07T17:44:49Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - Robust Bayesian Inference for Discrete Outcomes with the Total Variation
Distance [5.139874302398955]
Models of discrete-valued outcomes are easily misspecified if the data exhibit zero-inflation, overdispersion or contamination.
Here, we introduce a robust discrepancy-based Bayesian approach using the Total Variation Distance (TVD)
We empirically demonstrate that our approach is robust and significantly improves predictive performance on a range of simulated and real world data.
arXiv Detail & Related papers (2020-10-26T09:53:06Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.