Confidence-Aware Learning for Deep Neural Networks
- URL: http://arxiv.org/abs/2007.01458v3
- Date: Thu, 13 Aug 2020 03:16:37 GMT
- Title: Confidence-Aware Learning for Deep Neural Networks
- Authors: Jooyoung Moon, Jihyo Kim, Younghak Shin, Sangheum Hwang
- Abstract summary: We propose a method of training deep neural networks with a novel loss function, named Correctness Ranking Loss.
It regularizes class probabilities explicitly to be better confidence estimates in terms of ordinal ranking according to confidence.
It has almost the same computational costs for training as conventional deep classifiers and outputs reliable predictions by a single inference.
- Score: 4.9812879456945
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the power of deep neural networks for a wide range of tasks, an
overconfident prediction issue has limited their practical use in many
safety-critical applications. Many recent works have been proposed to mitigate
this issue, but most of them require either additional computational costs in
training and/or inference phases or customized architectures to output
confidence estimates separately. In this paper, we propose a method of training
deep neural networks with a novel loss function, named Correctness Ranking
Loss, which regularizes class probabilities explicitly to be better confidence
estimates in terms of ordinal ranking according to confidence. The proposed
method is easy to implement and can be applied to the existing architectures
without any modification. Also, it has almost the same computational costs for
training as conventional deep classifiers and outputs reliable predictions by a
single inference. Extensive experimental results on classification benchmark
datasets indicate that the proposed method helps networks to produce
well-ranked confidence estimates. We also demonstrate that it is effective for
the tasks closely related to confidence estimation, out-of-distribution
detection and active learning.
Related papers
- Likelihood-ratio-based confidence intervals for neural networks [1.4610038284393165]
This paper introduces a first implementation of a novel likelihood-ratio-based approach for constructing confidence intervals for neural networks.
While the current implementation of the method is prohibitively expensive for many deep-learning applications, the high cost may already be justified in specific fields like medical predictions or astrophysics.
This work highlights the significant potential of a likelihood-ratio-based uncertainty estimate and establishes a promising avenue for future research.
arXiv Detail & Related papers (2023-08-04T09:34:48Z) - Efficient Uncertainty Quantification and Reduction for
Over-Parameterized Neural Networks [23.7125322065694]
Uncertainty quantification (UQ) is important for reliability assessment and enhancement of machine learning models.
We create statistically guaranteed schemes to principally emphcharacterize, and emphremove, the uncertainty of over- parameterized neural networks.
In particular, our approach, based on what we call a procedural-noise-correcting (PNC) predictor, removes the procedural uncertainty by using only emphone auxiliary network that is trained on a suitably labeled dataset.
arXiv Detail & Related papers (2023-06-09T05:15:53Z) - Uncertainty Estimation by Fisher Information-based Evidential Deep
Learning [61.94125052118442]
Uncertainty estimation is a key factor that makes deep learning reliable in practical applications.
We propose a novel method, Fisher Information-based Evidential Deep Learning ($mathcalI$-EDL)
In particular, we introduce Fisher Information Matrix (FIM) to measure the informativeness of evidence carried by each sample, according to which we can dynamically reweight the objective loss terms to make the network more focused on the representation learning of uncertain classes.
arXiv Detail & Related papers (2023-03-03T16:12:59Z) - Towards Robust Deep Learning using Entropic Losses [0.0]
This thesis tackles the defiant out-of-distribution detection task by proposing novel loss functions and detection scores.
We also deal with this robustness-related task, which evaluates how realistic the probabilities presented by the deep neural network are.
arXiv Detail & Related papers (2022-08-06T18:52:39Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - Increasing the Confidence of Deep Neural Networks by Coverage Analysis [71.57324258813674]
This paper presents a lightweight monitoring architecture based on coverage paradigms to enhance the model against different unsafe inputs.
Experimental results show that the proposed approach is effective in detecting both powerful adversarial examples and out-of-distribution inputs.
arXiv Detail & Related papers (2021-01-28T16:38:26Z) - An evaluation of word-level confidence estimation for end-to-end
automatic speech recognition [70.61280174637913]
We investigate confidence estimation for end-to-end automatic speech recognition (ASR)
We provide an extensive benchmark of popular confidence methods on four well-known speech datasets.
Our results suggest a strong baseline can be obtained by scaling the logits by a learnt temperature.
arXiv Detail & Related papers (2021-01-14T09:51:59Z) - Multi-Loss Sub-Ensembles for Accurate Classification with Uncertainty
Estimation [1.2891210250935146]
We propose an efficient method for uncertainty estimation in deep neural networks (DNNs) achieving high accuracy.
We keep our inference time relatively low by leveraging the advantage proposed by the Deep-Sub-Ensembles method.
Our results show improved accuracy on the classification task and competitive results on several uncertainty measures.
arXiv Detail & Related papers (2020-10-05T10:59:11Z) - Cross Learning in Deep Q-Networks [82.20059754270302]
We propose a novel cross Q-learning algorithm, aim at alleviating the well-known overestimation problem in value-based reinforcement learning methods.
Our algorithm builds on double Q-learning, by maintaining a set of parallel models and estimate the Q-value based on a randomly selected network.
arXiv Detail & Related papers (2020-09-29T04:58:17Z) - Grasping Detection Network with Uncertainty Estimation for
Confidence-Driven Semi-Supervised Domain Adaptation [17.16216430459064]
This paper presents an approach enabling the easy domain adaptation through a novel grasping detection network with confidence-driven semi-supervised learning.
The proposed grasping detection network specially provides a prediction uncertainty estimation mechanism by leveraging on Feature Pyramid Network (FPN), and the mean-teacher semi-supervised learning utilizes such uncertainty information to emphasizing the consistency loss only for those unlabelled data with high confidence.
Our results show that the proposed network can achieve high success rate on the Cornell grasping dataset, and for domain adaptation with very limited data, the confidence-driven mean teacher outperforms the original mean teacher and direct training by more than 10% in evaluation
arXiv Detail & Related papers (2020-08-20T07:42:45Z) - Being Bayesian, Even Just a Bit, Fixes Overconfidence in ReLU Networks [65.24701908364383]
We show that a sufficient condition for a uncertainty on a ReLU network is "to be a bit Bayesian calibrated"
We further validate these findings empirically via various standard experiments using common deep ReLU networks and Laplace approximations.
arXiv Detail & Related papers (2020-02-24T08:52:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.