Local Temperature Scaling for Probability Calibration
- URL: http://arxiv.org/abs/2008.05105v2
- Date: Tue, 27 Jul 2021 10:30:12 GMT
- Title: Local Temperature Scaling for Probability Calibration
- Authors: Zhipeng Ding, Xu Han, Peirong Liu, Marc Niethammer
- Abstract summary: We propose a learning-based calibration method that focuses on semantic segmentation.
Specifically, we adopt a convolutional neural network to predict local temperature values for probability calibration.
Experiments on the COCO, CamVid, and LPBA40 datasets demonstrate improved calibration performance for a range of different metrics.
- Score: 22.069749881109992
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For semantic segmentation, label probabilities are often uncalibrated as they
are typically only the by-product of a segmentation task. Intersection over
Union (IoU) and Dice score are often used as criteria for segmentation success,
while metrics related to label probabilities are not often explored. However,
probability calibration approaches have been studied, which match probability
outputs with experimentally observed errors. These approaches mainly focus on
classification tasks, but not on semantic segmentation. Thus, we propose a
learning-based calibration method that focuses on multi-label semantic
segmentation. Specifically, we adopt a convolutional neural network to predict
local temperature values for probability calibration. One advantage of our
approach is that it does not change prediction accuracy, hence allowing for
calibration as a post-processing step. Experiments on the COCO, CamVid, and
LPBA40 datasets demonstrate improved calibration performance for a range of
different metrics. We also demonstrate the good performance of our method for
multi-atlas brain segmentation from magnetic resonance images.
Related papers
- SoftCVI: Contrastive variational inference with self-generated soft labels [2.5398014196797614]
Variational inference and Markov chain Monte Carlo methods are the predominant tools for this task.
We introduce Soft Contrastive Variational Inference (SoftCVI), which allows a family of variational objectives to be derived through a contrastive estimation framework.
We find that SoftCVI can be used to form objectives which are stable to train and mass-covering, frequently outperforming inference with other variational approaches.
arXiv Detail & Related papers (2024-07-22T14:54:12Z) - Calibration by Distribution Matching: Trainable Kernel Calibration
Metrics [56.629245030893685]
We introduce kernel-based calibration metrics that unify and generalize popular forms of calibration for both classification and regression.
These metrics admit differentiable sample estimates, making it easy to incorporate a calibration objective into empirical risk minimization.
We provide intuitive mechanisms to tailor calibration metrics to a decision task, and enforce accurate loss estimation and no regret decisions.
arXiv Detail & Related papers (2023-10-31T06:19:40Z) - Model Calibration in Dense Classification with Adaptive Label
Perturbation [44.62722402349157]
Existing dense binary classification models are prone to being over-confident.
We propose Adaptive Label Perturbation (ASLP) which learns a unique label perturbation level for each training image.
ASLP can significantly improve calibration degrees of dense binary classification models on both in-distribution and out-of-distribution data.
arXiv Detail & Related papers (2023-07-25T14:40:11Z) - Towards Better Certified Segmentation via Diffusion Models [62.21617614504225]
segmentation models can be vulnerable to adversarial perturbations, which hinders their use in critical-decision systems like healthcare or autonomous driving.
Recently, randomized smoothing has been proposed to certify segmentation predictions by adding Gaussian noise to the input to obtain theoretical guarantees.
In this paper, we address the problem of certifying segmentation prediction using a combination of randomized smoothing and diffusion models.
arXiv Detail & Related papers (2023-06-16T16:30:39Z) - On Calibrating Semantic Segmentation Models: Analyses and An Algorithm [51.85289816613351]
We study the problem of semantic segmentation calibration.
Model capacity, crop size, multi-scale testing, and prediction correctness have impact on calibration.
We propose a simple, unifying, and effective approach, namely selective scaling.
arXiv Detail & Related papers (2022-12-22T22:05:16Z) - Heatmap Regression for Lesion Detection using Pointwise Annotations [3.6513059119482145]
In this paper, we present a lesion detection method which relies only on point labels.
Our model, which is trained via heatmap regression, can detect a variable number of lesions in a probabilistic manner.
Experimental results on Gad lesion detection show our point-based method performs competitively compared to training on expensive segmentation labels.
arXiv Detail & Related papers (2022-08-11T17:26:09Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Unsupervised Calibration under Covariate Shift [92.02278658443166]
We introduce the problem of calibration under domain shift and propose an importance sampling based approach to address it.
We evaluate and discuss the efficacy of our method on both real-world datasets and synthetic datasets.
arXiv Detail & Related papers (2020-06-29T21:50:07Z) - Calibration of Neural Networks using Splines [51.42640515410253]
Measuring calibration error amounts to comparing two empirical distributions.
We introduce a binning-free calibration measure inspired by the classical Kolmogorov-Smirnov (KS) statistical test.
Our method consistently outperforms existing methods on KS error as well as other commonly used calibration measures.
arXiv Detail & Related papers (2020-06-23T07:18:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.