Automatic Open-World Reliability Assessment
- URL: http://arxiv.org/abs/2011.05506v2
- Date: Mon, 14 Dec 2020 01:35:18 GMT
- Title: Automatic Open-World Reliability Assessment
- Authors: Mohsen Jafarzadeh, Touqeer Ahmad, Akshay Raj Dhamija, Chunchun Li,
Steve Cruz, Terrance E. Boult
- Abstract summary: Image classification in the open-world must handle out-of-distribution (OOD) images.
We formalize the open-world recognition reliability problem and propose multiple automatic reliability assessment policies.
- Score: 11.380522815465985
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image classification in the open-world must handle out-of-distribution (OOD)
images. Systems should ideally reject OOD images, or they will map atop of
known classes and reduce reliability. Using open-set classifiers that can
reject OOD inputs can help. However, optimal accuracy of open-set classifiers
depend on the frequency of OOD data. Thus, for either standard or open-set
classifiers, it is important to be able to determine when the world changes and
increasing OOD inputs will result in reduced system reliability. However,
during operations, we cannot directly assess accuracy as there are no labels.
Thus, the reliability assessment of these classifiers must be done by human
operators, made more complex because networks are not 100% accurate, so some
failures are to be expected. To automate this process, herein, we formalize the
open-world recognition reliability problem and propose multiple automatic
reliability assessment policies to address this new problem using only the
distribution of reported scores/probability data. The distributional algorithms
can be applied to both classic classifiers with SoftMax as well as the
open-world Extreme Value Machine (EVM) to provide automated reliability
assessment. We show that all of the new algorithms significantly outperform
detection using the mean of SoftMax.
Related papers
- What If the Input is Expanded in OOD Detection? [77.37433624869857]
Out-of-distribution (OOD) detection aims to identify OOD inputs from unknown classes.
Various scoring functions are proposed to distinguish it from in-distribution (ID) data.
We introduce a novel perspective, i.e., employing different common corruptions on the input space.
arXiv Detail & Related papers (2024-10-24T06:47:28Z) - How Does Unlabeled Data Provably Help Out-of-Distribution Detection? [63.41681272937562]
Unlabeled in-the-wild data is non-trivial due to the heterogeneity of both in-distribution (ID) and out-of-distribution (OOD) data.
This paper introduces a new learning framework SAL (Separate And Learn) that offers both strong theoretical guarantees and empirical effectiveness.
arXiv Detail & Related papers (2024-02-05T20:36:33Z) - Binary Classification with Confidence Difference [100.08818204756093]
This paper delves into a novel weakly supervised binary classification problem called confidence-difference (ConfDiff) classification.
We propose a risk-consistent approach to tackle this problem and show that the estimation error bound the optimal convergence rate.
We also introduce a risk correction approach to mitigate overfitting problems, whose consistency and convergence rate are also proven.
arXiv Detail & Related papers (2023-10-09T11:44:50Z) - Free Lunch for Generating Effective Outlier Supervision [46.37464572099351]
We propose an ultra-effective method to generate near-realistic outlier supervision.
Our proposed textttBayesAug significantly reduces the false positive rate over 12.50% compared with the previous schemes.
arXiv Detail & Related papers (2023-01-17T01:46:45Z) - Accurate and Reliable Methods for 5G UAV Jamming Identification With
Calibrated Uncertainty [3.4208659698673127]
Only increasing accuracy without considering uncertainty may negatively impact Deep Neural Network (DNN) decision-making.
This paper proposes five combined preprocessing and post-processing methods for time-series binary classification problems.
arXiv Detail & Related papers (2022-11-05T15:04:45Z) - Evaluating Uncertainty Calibration for Open-Set Recognition [5.8022510096020525]
Deep neural networks (DNNs) suffer from providing overconfident probabilities on out-of-distribution (OOD) data.
We evaluate popular calibration techniques for open-set conditions in a way that is distinctly different from the conventional evaluation of calibration methods on OOD data.
arXiv Detail & Related papers (2022-05-15T02:08:35Z) - Provably Robust Detection of Out-of-distribution Data (almost) for free [124.14121487542613]
Deep neural networks are known to produce highly overconfident predictions on out-of-distribution (OOD) data.
In this paper we propose a novel method where from first principles we combine a certifiable OOD detector with a standard classifier into an OOD aware classifier.
In this way we achieve the best of two worlds: certifiably adversarially robust OOD detection, even for OOD samples close to the in-distribution, without loss in prediction accuracy and close to state-of-the-art OOD detection performance for non-manipulated OOD data.
arXiv Detail & Related papers (2021-06-08T11:40:49Z) - Out-of-Distribution Detection for Automotive Perception [58.34808836642603]
Neural networks (NNs) are widely used for object classification in autonomous driving.
NNs can fail on input data not well represented by the training dataset, known as out-of-distribution (OOD) data.
This paper presents a method for determining whether inputs are OOD, which does not require OOD data during training and does not increase the computational cost of inference.
arXiv Detail & Related papers (2020-11-03T01:46:35Z) - Detecting Misclassification Errors in Neural Networks with a Gaussian
Process Model [20.948038514886377]
This paper presents a new framework that produces a quantitative metric for detecting misclassification errors.
The framework, RED, builds an error detector on top of the base classifier and estimates uncertainty of the detection scores using Gaussian Processes.
arXiv Detail & Related papers (2020-10-05T15:01:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.