Fine-grained Uncertainty Modeling in Neural Networks
- URL: http://arxiv.org/abs/2002.04205v1
- Date: Tue, 11 Feb 2020 05:06:25 GMT
- Title: Fine-grained Uncertainty Modeling in Neural Networks
- Authors: Rahul Soni, Naresh Shah, Jimmy D. Moore
- Abstract summary: We present a novel method to detect out-of-distribution points in a Neural Network.
Our method corrects overconfident NN decisions, detects outlier points and learns to say I don't know'' when uncertain about a critical point between the top two predictions.
As a positive side effect, our method helps to prevent adversarial attacks without requiring any additional training.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing uncertainty modeling approaches try to detect an out-of-distribution
point from the in-distribution dataset. We extend this argument to detect
finer-grained uncertainty that distinguishes between (a). certain points, (b).
uncertain points but within the data distribution, and (c). out-of-distribution
points. Our method corrects overconfident NN decisions, detects outlier points
and learns to say ``I don't know'' when uncertain about a critical point
between the top two predictions. In addition, we provide a mechanism to
quantify class distributions overlap in the decision manifold and investigate
its implications in model interpretability.
Our method is two-step: in the first step, the proposed method builds a class
distribution using Kernel Activation Vectors (kav) extracted from the Network.
In the second step, the algorithm determines the confidence of a test point by
a hierarchical decision rule based on the chi-squared distribution of squared
Mahalanobis distances.
Our method sits on top of a given Neural Network, requires a single scan of
training data to estimate class distribution statistics, and is highly scalable
to deep networks and wider pre-softmax layer. As a positive side effect, our
method helps to prevent adversarial attacks without requiring any additional
training. It is directly achieved when the Softmax layer is substituted by our
robust uncertainty layer at the evaluation phase.
Related papers
- Combining Statistical Depth and Fermat Distance for Uncertainty Quantification [3.3975558777609915]
We measure the Out-of-domain uncertainty in the prediction of Neural Networks using a statistical notion called Lens Depth'' (LD) combined with Fermat Distance.
The proposed method gives excellent qualitative result on toy datasets and can give competitive or better uncertainty estimation on standard deep learning datasets.
arXiv Detail & Related papers (2024-04-12T13:54:21Z) - Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment [87.8301166955305]
We propose a novel bi-discriminator domain adversarial neural network with class-level gradient alignment.
BACG resorts to gradient signals and second-order probability estimation for better alignment of domain distributions.
In addition, inspired by contrastive learning, we develop a memory bank-based variant, i.e. Fast-BACG, which can greatly shorten the training process.
arXiv Detail & Related papers (2023-10-21T09:53:17Z) - Improved uncertainty quantification for neural networks with Bayesian
last layer [0.0]
Uncertainty quantification is an important task in machine learning.
We present a reformulation of the log-marginal likelihood of a NN with BLL which allows for efficient training using backpropagation.
arXiv Detail & Related papers (2023-02-21T20:23:56Z) - NUQ: Nonparametric Uncertainty Quantification for Deterministic Neural
Networks [151.03112356092575]
We show the principled way to measure the uncertainty of predictions for a classifier based on Nadaraya-Watson's nonparametric estimate of the conditional label distribution.
We demonstrate the strong performance of the method in uncertainty estimation tasks on a variety of real-world image datasets.
arXiv Detail & Related papers (2022-02-07T12:30:45Z) - Dense Uncertainty Estimation [62.23555922631451]
In this paper, we investigate neural networks and uncertainty estimation techniques to achieve both accurate deterministic prediction and reliable uncertainty estimation.
We work on two types of uncertainty estimations solutions, namely ensemble based methods and generative model based methods, and explain their pros and cons while using them in fully/semi/weakly-supervised framework.
arXiv Detail & Related papers (2021-10-13T01:23:48Z) - Learning Calibrated Uncertainties for Domain Shift: A Distributionally
Robust Learning Approach [150.8920602230832]
We propose a framework for learning calibrated uncertainties under domain shifts.
In particular, the density ratio estimation reflects the closeness of a target (test) sample to the source (training) distribution.
We show that our proposed method generates calibrated uncertainties that benefit downstream tasks.
arXiv Detail & Related papers (2020-10-08T02:10:54Z) - Revisiting One-vs-All Classifiers for Predictive Uncertainty and
Out-of-Distribution Detection in Neural Networks [22.34227625637843]
We investigate how the parametrization of the probabilities in discriminative classifiers affects the uncertainty estimates.
We show that one-vs-all formulations can improve calibration on image classification tasks.
arXiv Detail & Related papers (2020-07-10T01:55:02Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Uncertainty Estimation Using a Single Deep Deterministic Neural Network [66.26231423824089]
We propose a method for training a deterministic deep model that can find and reject out of distribution data points at test time with a single forward pass.
We scale training in these with a novel loss function and centroid updating scheme and match the accuracy of softmax models.
arXiv Detail & Related papers (2020-03-04T12:27:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.