Intra Order-preserving Functions for Calibration of Multi-Class Neural
Networks
- URL: http://arxiv.org/abs/2003.06820v2
- Date: Fri, 23 Oct 2020 06:59:28 GMT
- Title: Intra Order-preserving Functions for Calibration of Multi-Class Neural
Networks
- Authors: Amir Rahimi, Amirreza Shaban, Ching-An Cheng, Richard Hartley, Byron
Boots
- Abstract summary: A common approach is to learn a post-hoc calibration function that transforms the output of the original network into calibrated confidence scores.
Previous post-hoc calibration techniques work only with simple calibration functions.
We propose a new neural network architecture that represents a class of intra order-preserving functions.
- Score: 54.23874144090228
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Predicting calibrated confidence scores for multi-class deep networks is
important for avoiding rare but costly mistakes. A common approach is to learn
a post-hoc calibration function that transforms the output of the original
network into calibrated confidence scores while maintaining the network's
accuracy. However, previous post-hoc calibration techniques work only with
simple calibration functions, potentially lacking sufficient representation to
calibrate the complex function landscape of deep networks. In this work, we aim
to learn general post-hoc calibration functions that can preserve the top-k
predictions of any deep network. We call this family of functions intra
order-preserving functions. We propose a new neural network architecture that
represents a class of intra order-preserving functions by combining common
neural network components. Additionally, we introduce order-invariant and
diagonal sub-families, which can act as regularization for better
generalization when the training data size is small. We show the effectiveness
of the proposed method across a wide range of datasets and classifiers. Our
method outperforms state-of-the-art post-hoc calibration methods, namely
temperature scaling and Dirichlet calibration, in several evaluation metrics
for the task.
Related papers
- Feature Clipping for Uncertainty Calibration [24.465567005078135]
Modern deep neural networks (DNNs) often suffer from overconfidence, leading to miscalibration.
We propose a novel post-hoc calibration method called feature clipping (FC) to address this issue.
FC involves clipping feature values to a specified threshold, effectively increasing entropy in high calibration error samples.
arXiv Detail & Related papers (2024-10-16T06:44:35Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - Multi-Head Multi-Loss Model Calibration [13.841172927454204]
We introduce a form of simplified ensembling that bypasses the costly training and inference of deep ensembles.
Specifically, each head is trained to minimize a weighted Cross-Entropy loss, but the weights are different among the different branches.
We show that the resulting averaged predictions can achieve excellent calibration without sacrificing accuracy in two challenging datasets.
arXiv Detail & Related papers (2023-03-02T09:32:32Z) - Neural Clamping: Joint Input Perturbation and Temperature Scaling for Neural Network Calibration [62.4971588282174]
We propose a new post-processing calibration method called Neural Clamping.
Our empirical results show that Neural Clamping significantly outperforms state-of-the-art post-processing calibration methods.
arXiv Detail & Related papers (2022-09-23T14:18:39Z) - On Calibration of Graph Neural Networks for Node Classification [29.738179864433445]
Graph neural networks learn entity and edge embeddings for tasks such as node classification and link prediction.
These models achieve good performance with respect to accuracy, but the confidence scores associated with the predictions might not be calibrated.
We propose a topology-aware calibration method that takes the neighboring nodes into account and yields improved calibration.
arXiv Detail & Related papers (2022-06-03T13:48:10Z) - Parameterized Temperature Scaling for Boosting the Expressive Power in
Post-Hoc Uncertainty Calibration [57.568461777747515]
We introduce a novel calibration method, Parametrized Temperature Scaling (PTS)
We demonstrate that the performance of accuracy-preserving state-of-the-art post-hoc calibrators is limited by their intrinsic expressive power.
We show with extensive experiments that our novel accuracy-preserving approach consistently outperforms existing algorithms across a large number of model architectures, datasets and metrics.
arXiv Detail & Related papers (2021-02-24T10:18:30Z) - Improved Trainable Calibration Method for Neural Networks on Medical
Imaging Classification [17.941506832422192]
Empirically, neural networks are often miscalibrated and overconfident in their predictions.
We propose a novel calibration approach that maintains the overall classification accuracy while significantly improving model calibration.
arXiv Detail & Related papers (2020-09-09T01:25:53Z) - Post-hoc Calibration of Neural Networks by g-Layers [51.42640515410253]
In recent years, there is a surge of research on neural network calibration.
It is known that minimizing Negative Log-Likelihood (NLL) will lead to a calibrated network on the training set if the global optimum is attained.
We prove that even though the base network ($f$) does not lead to the global optimum of NLL, by adding additional layers ($g$) and minimizing NLL by optimizing the parameters of $g$ one can obtain a calibrated network.
arXiv Detail & Related papers (2020-06-23T07:55:10Z) - Calibrating Deep Neural Networks using Focal Loss [77.92765139898906]
Miscalibration is a mismatch between a model's confidence and its correctness.
We show that focal loss allows us to learn models that are already very well calibrated.
We show that our approach achieves state-of-the-art calibration without compromising on accuracy in almost all cases.
arXiv Detail & Related papers (2020-02-21T17:35:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.