GSC Loss: A Gaussian Score Calibrating Loss for Deep Learning
- URL: http://arxiv.org/abs/2203.00833v1
- Date: Wed, 2 Mar 2022 02:52:23 GMT
- Title: GSC Loss: A Gaussian Score Calibrating Loss for Deep Learning
- Authors: Qingsong Zhao, Shuguang Dou, Xiaopeng Ji, Xinyang Jiang, Cairong Zhao,
Yin Wang
- Abstract summary: We propose a general Gaussian Score Calibrating (GSC) loss to calibrate the predicted scores produced by the deep neural networks (DNN)
Extensive experiments on over 10 benchmark datasets demonstrate that the proposed GSC loss can yield consistent and significant performance boosts in a variety of visual tasks.
- Score: 16.260520216972854
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Cross entropy (CE) loss integrated with softmax is an orthodox component in
most classification-based frameworks, but it fails to obtain an accurate
probability distribution of predicted scores that is critical for further
decision-making of poor-classified samples. The prediction score calibration
provides a solution to learn the distribution of predicted scores which can
explicitly make the model obtain a discriminative representation. Considering
the entropy function can be utilized to measure the uncertainty of predicted
scores. But, the gradient variation of it is not in line with the expectations
of model optimization. To this end, we proposed a general Gaussian Score
Calibrating (GSC) loss to calibrate the predicted scores produced by the deep
neural networks (DNN). Extensive experiments on over 10 benchmark datasets
demonstrate that the proposed GSC loss can yield consistent and significant
performance boosts in a variety of visual tasks. Notably, our label-independent
GSC loss can be embedded into common improved methods based on the CE loss
easily.
Related papers
- Calibrating Deep Neural Network using Euclidean Distance [5.675312975435121]
In machine learning, Focal Loss is commonly used to reduce misclassification rates by emphasizing hard-to-classify samples.
High calibration error indicates a misalignment between predicted probabilities and actual outcomes, affecting model reliability.
This research introduces a novel loss function called Focal Loss (FCL), designed to improve probability calibration while retaining the advantages of Focal Loss in handling difficult samples.
arXiv Detail & Related papers (2024-10-23T23:06:50Z) - Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Model Calibration in Dense Classification with Adaptive Label
Perturbation [44.62722402349157]
Existing dense binary classification models are prone to being over-confident.
We propose Adaptive Label Perturbation (ASLP) which learns a unique label perturbation level for each training image.
ASLP can significantly improve calibration degrees of dense binary classification models on both in-distribution and out-of-distribution data.
arXiv Detail & Related papers (2023-07-25T14:40:11Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Boosting Differentiable Causal Discovery via Adaptive Sample Reweighting [62.23057729112182]
Differentiable score-based causal discovery methods learn a directed acyclic graph from observational data.
We propose a model-agnostic framework to boost causal discovery performance by dynamically learning the adaptive weights for the Reweighted Score function, ReScore.
arXiv Detail & Related papers (2023-03-06T14:49:59Z) - Beyond calibration: estimating the grouping loss of modern neural
networks [68.8204255655161]
Proper scoring rule theory shows that given the calibration loss, the missing piece to characterize individual errors is the grouping loss.
We show that modern neural network architectures in vision and NLP exhibit grouping loss, notably in distribution shifts settings.
arXiv Detail & Related papers (2022-10-28T07:04:20Z) - Fine-grained Classification of Solder Joints with {\alpha}-skew
Jensen-Shannon Divergence [0.0]
We show that solders have low feature diversity, and that the solder joint inspection can be carried out as a fine-grained image classification task.
To improve the fine-grained classification accuracy, penalizing confident model predictions by maximizing entropy was found useful in the literature.
We show that the proposed approach achieves the highest F1-score and competitive accuracy for different models in the finegrained solder joint classification task.
arXiv Detail & Related papers (2022-09-20T17:06:51Z) - EXACT: How to Train Your Accuracy [6.144680854063938]
We propose a new optimization framework by introducing ascentity to a model's output and optimizing expected accuracy.
Experiments on linear models and deep image classification show that the proposed optimization method is a powerful alternative to widely used classification losses.
arXiv Detail & Related papers (2022-05-19T15:13:00Z) - The Devil is in the Margin: Margin-based Label Smoothing for Network
Calibration [21.63888208442176]
In spite of the dominant performances of deep neural networks, recent works have shown that they are poorly calibrated.
We provide a unifying constrained-optimization perspective of current state-of-the-art calibration losses.
We propose a simple and flexible generalization based on inequality constraints, which imposes a controllable margin on logit distances.
arXiv Detail & Related papers (2021-11-30T14:21:47Z) - Shaping Deep Feature Space towards Gaussian Mixture for Visual
Classification [74.48695037007306]
We propose a Gaussian mixture (GM) loss function for deep neural networks for visual classification.
With a classification margin and a likelihood regularization, the GM loss facilitates both high classification performance and accurate modeling of the feature distribution.
The proposed model can be implemented easily and efficiently without using extra trainable parameters.
arXiv Detail & Related papers (2020-11-18T03:32:27Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.