Plots of the cumulative differences between observed and expected values
of ordered Bernoulli variates
- URL: http://arxiv.org/abs/2006.02504v3
- Date: Fri, 17 Jul 2020 01:49:14 GMT
- Title: Plots of the cumulative differences between observed and expected values
of ordered Bernoulli variates
- Authors: Mark Tygert
- Abstract summary: "Reliability diagrams" (also known as "calibration plots") help detect and diagnose significant discrepancies between predictions and outcomes.
The canonical reliability diagrams are based on histogramming the observed and expected values of the predictions.
Several variants of the standard reliability diagrams propose to replace the hard histogram binning with soft kernel density estimation.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many predictions are probabilistic in nature; for example, a prediction could
be for precipitation tomorrow, but with only a 30 percent chance. Given both
the predictions and the actual outcomes, "reliability diagrams" (also known as
"calibration plots") help detect and diagnose statistically significant
discrepancies between the predictions and the outcomes. The canonical
reliability diagrams are based on histogramming the observed and expected
values of the predictions; several variants of the standard reliability
diagrams propose to replace the hard histogram binning with soft kernel density
estimation using smooth convolutional kernels of widths similar to the widths
of the bins. In all cases, an important question naturally arises: which widths
are best (or are multiple plots with different widths better)? Rather than
answering this question, plots of the cumulative differences between the
observed and expected values largely avoid the question, by displaying
miscalibration directly as the slopes of secant lines for the graphs. Slope is
easy to perceive with quantitative precision even when the constant offsets of
the secant lines are irrelevant. There is no need to bin or perform kernel
density estimation with a somewhat arbitrary kernel.
Related papers
- Semiparametric conformal prediction [79.6147286161434]
Risk-sensitive applications require well-calibrated prediction sets over multiple, potentially correlated target variables.
We treat the scores as random vectors and aim to construct the prediction set accounting for their joint correlation structure.
We report desired coverage and competitive efficiency on a range of real-world regression problems.
arXiv Detail & Related papers (2024-11-04T14:29:02Z) - SoftCVI: Contrastive variational inference with self-generated soft labels [2.5398014196797614]
Variational inference and Markov chain Monte Carlo methods are the predominant tools for this task.
We introduce Soft Contrastive Variational Inference (SoftCVI), which allows a family of variational objectives to be derived through a contrastive estimation framework.
We find that SoftCVI can be used to form objectives which are stable to train and mass-covering, frequently outperforming inference with other variational approaches.
arXiv Detail & Related papers (2024-07-22T14:54:12Z) - Awareness of uncertainty in classification using a multivariate model and multi-views [1.3048920509133808]
The proposed model regularizes uncertain predictions, and trains to calculate both the predictions and their uncertainty estimations.
Given the multi-view predictions together with their uncertainties and confidences, we proposed several methods to calculate final predictions.
The proposed methodology was tested using CIFAR-10 dataset with clean and noisy labels.
arXiv Detail & Related papers (2024-04-16T06:40:51Z) - Confidence and Dispersity Speak: Characterising Prediction Matrix for
Unsupervised Accuracy Estimation [51.809741427975105]
This work aims to assess how well a model performs under distribution shifts without using labels.
We use the nuclear norm that has been shown to be effective in characterizing both properties.
We show that the nuclear norm is more accurate and robust in accuracy than existing methods.
arXiv Detail & Related papers (2023-02-02T13:30:48Z) - Nonparametric Probabilistic Regression with Coarse Learners [1.8275108630751844]
We show that we can compute precise conditional densities with minimal assumptions on the shape or form of the density.
We demonstrate this approach on a variety of datasets and show competitive performance, particularly on larger datasets.
arXiv Detail & Related papers (2022-10-28T16:25:26Z) - Precise Learning Curves and Higher-Order Scaling Limits for Dot Product
Kernel Regression [41.48538038768993]
We focus on the problem of kernel ridge regression for dot-product kernels.
We observe a peak in the learning curve whenever $m approx dr/r!$ for any integer $r$, leading to multiple sample-wise descent and nontrivial behavior at multiple scales.
arXiv Detail & Related papers (2022-05-30T04:21:31Z) - Metrics of calibration for probabilistic predictions [0.0]
"Reliability diagrams" help detect and diagnose statistically significant discrepancies -- so-called "miscalibration"
The canonical reliability diagrams histogram the observed and expected values of the predictions.
But, which widths of bins or kernels are best?
Slope is easy to perceive with quantitative precision, even when the constant offsets of the secant lines are irrelevant.
arXiv Detail & Related papers (2022-05-19T16:38:24Z) - Multivariate Probabilistic Regression with Natural Gradient Boosting [63.58097881421937]
We propose a Natural Gradient Boosting (NGBoost) approach based on nonparametrically modeling the conditional parameters of the multivariate predictive distribution.
Our method is robust, works out-of-the-box without extensive tuning, is modular with respect to the assumed target distribution, and performs competitively in comparison to existing approaches.
arXiv Detail & Related papers (2021-06-07T17:44:49Z) - Learning Accurate Dense Correspondences and When to Trust Them [161.76275845530964]
We aim to estimate a dense flow field relating two images, coupled with a robust pixel-wise confidence map.
We develop a flexible probabilistic approach that jointly learns the flow prediction and its uncertainty.
Our approach obtains state-of-the-art results on challenging geometric matching and optical flow datasets.
arXiv Detail & Related papers (2021-01-05T18:54:11Z) - Learning Disentangled Representations with Latent Variation
Predictability [102.4163768995288]
This paper defines the variation predictability of latent disentangled representations.
Within an adversarial generation process, we encourage variation predictability by maximizing the mutual information between latent variations and corresponding image pairs.
We develop an evaluation metric that does not rely on the ground-truth generative factors to measure the disentanglement of latent representations.
arXiv Detail & Related papers (2020-07-25T08:54:26Z) - Bayesian Deep Learning and a Probabilistic Perspective of Generalization [56.69671152009899]
We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization.
We also propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction.
arXiv Detail & Related papers (2020-02-20T15:13:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.