Meta-Calibration: Learning of Model Calibration Using Differentiable
Expected Calibration Error
- URL: http://arxiv.org/abs/2106.09613v3
- Date: Fri, 25 Aug 2023 19:18:31 GMT
- Title: Meta-Calibration: Learning of Model Calibration Using Differentiable
Expected Calibration Error
- Authors: Ondrej Bohdal, Yongxin Yang, Timothy Hospedales
- Abstract summary: We introduce a new differentiable surrogate for expected calibration error (DECE) that allows calibration quality to be directly optimised.
We also propose a meta-learning framework that uses DECE to optimise for validation set calibration.
- Score: 46.12703434199988
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Calibration of neural networks is a topical problem that is becoming more and
more important as neural networks increasingly underpin real-world
applications. The problem is especially noticeable when using modern neural
networks, for which there is a significant difference between the confidence of
the model and the probability of correct prediction. Various strategies have
been proposed to improve calibration, yet accurate calibration remains
challenging. We propose a novel framework with two contributions: introducing a
new differentiable surrogate for expected calibration error (DECE) that allows
calibration quality to be directly optimised, and a meta-learning framework
that uses DECE to optimise for validation set calibration with respect to model
hyper-parameters. The results show that we achieve competitive performance with
existing calibration approaches. Our framework opens up a new avenue and
toolset for tackling calibration, which we believe will inspire further work on
this important challenge.
Related papers
- Feature Clipping for Uncertainty Calibration [24.465567005078135]
Modern deep neural networks (DNNs) often suffer from overconfidence, leading to miscalibration.
We propose a novel post-hoc calibration method called feature clipping (FC) to address this issue.
FC involves clipping feature values to a specified threshold, effectively increasing entropy in high calibration error samples.
arXiv Detail & Related papers (2024-10-16T06:44:35Z) - Towards Unbiased Calibration using Meta-Regularization [6.440598446802981]
We propose to learn better-calibrated models via meta-regularization, which has two components.
We evaluate the effectiveness of the proposed approach in regularizing neural networks towards improved and unbiased calibration on three computer vision datasets.
arXiv Detail & Related papers (2023-03-27T10:00:50Z) - Calibration of Neural Networks [77.34726150561087]
This paper presents a survey of confidence calibration problems in the context of neural networks.
We analyze problem statement, calibration definitions, and different approaches to evaluation.
Empirical experiments cover various datasets and models, comparing calibration methods according to different criteria.
arXiv Detail & Related papers (2023-03-19T20:27:51Z) - On Calibrating Semantic Segmentation Models: Analyses and An Algorithm [51.85289816613351]
We study the problem of semantic segmentation calibration.
Model capacity, crop size, multi-scale testing, and prediction correctness have impact on calibration.
We propose a simple, unifying, and effective approach, namely selective scaling.
arXiv Detail & Related papers (2022-12-22T22:05:16Z) - Revisiting Calibration for Question Answering [16.54743762235555]
We argue that the traditional evaluation of calibration does not reflect usefulness of the model confidence.
We propose a new calibration metric, MacroCE, that better captures whether the model assigns low confidence to wrong predictions and high confidence to correct predictions.
arXiv Detail & Related papers (2022-05-25T05:49:56Z) - On the Dark Side of Calibration for Modern Neural Networks [65.83956184145477]
We show the breakdown of expected calibration error (ECE) into predicted confidence and refinement.
We highlight that regularisation based calibration only focuses on naively reducing a model's confidence.
We find that many calibration approaches with the likes of label smoothing, mixup etc. lower the utility of a DNN by degrading its refinement.
arXiv Detail & Related papers (2021-06-17T11:04:14Z) - Post-hoc Calibration of Neural Networks by g-Layers [51.42640515410253]
In recent years, there is a surge of research on neural network calibration.
It is known that minimizing Negative Log-Likelihood (NLL) will lead to a calibrated network on the training set if the global optimum is attained.
We prove that even though the base network ($f$) does not lead to the global optimum of NLL, by adding additional layers ($g$) and minimizing NLL by optimizing the parameters of $g$ one can obtain a calibrated network.
arXiv Detail & Related papers (2020-06-23T07:55:10Z) - Calibrating Deep Neural Networks using Focal Loss [77.92765139898906]
Miscalibration is a mismatch between a model's confidence and its correctness.
We show that focal loss allows us to learn models that are already very well calibrated.
We show that our approach achieves state-of-the-art calibration without compromising on accuracy in almost all cases.
arXiv Detail & Related papers (2020-02-21T17:35:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.