Rethinking Out-of-Distribution Detection From a Human-Centric
Perspective
- URL: http://arxiv.org/abs/2211.16778v1
- Date: Wed, 30 Nov 2022 06:34:50 GMT
- Title: Rethinking Out-of-Distribution Detection From a Human-Centric
Perspective
- Authors: Yao Zhu, Yuefeng Chen, Xiaodan Li, Rong Zhang, Hui Xue, Xiang Tian,
Rongxin Jiang, Bolun Zheng, Yaowu Chen
- Abstract summary: Out-Of-Distribution (OOD) detection aims to ensure the reliability and safety of deep neural networks (DNNs) in real-world scenarios.
We propose a human-centric evaluation and conduct extensive experiments on 45 classifiers and 8 test datasets.
We find that the simple baseline OOD detection method can achieve comparable and even better performance than the recently proposed methods.
- Score: 22.834986963880482
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Out-Of-Distribution (OOD) detection has received broad attention over the
years, aiming to ensure the reliability and safety of deep neural networks
(DNNs) in real-world scenarios by rejecting incorrect predictions. However, we
notice a discrepancy between the conventional evaluation vs. the essential
purpose of OOD detection. On the one hand, the conventional evaluation
exclusively considers risks caused by label-space distribution shifts while
ignoring the risks from input-space distribution shifts. On the other hand, the
conventional evaluation reward detection methods for not rejecting the
misclassified image in the validation dataset. However, the misclassified image
can also cause risks and should be rejected. We appeal to rethink OOD detection
from a human-centric perspective, that a proper detection method should reject
the case that the deep model's prediction mismatches the human expectations and
adopt the case that the deep model's prediction meets the human expectations.
We propose a human-centric evaluation and conduct extensive experiments on 45
classifiers and 8 test datasets. We find that the simple baseline OOD detection
method can achieve comparable and even better performance than the recently
proposed methods, which means that the development in OOD detection in the past
years may be overestimated. Additionally, our experiments demonstrate that
model selection is non-trivial for OOD detection and should be considered as an
integral of the proposed method, which differs from the claim in existing works
that proposed methods are universal across different models.
Related papers
- The Best of Both Worlds: On the Dilemma of Out-of-distribution Detection [75.65876949930258]
Out-of-distribution (OOD) detection is essential for model trustworthiness.
We show that the superior OOD detection performance of state-of-the-art methods is achieved by secretly sacrificing the OOD generalization ability.
arXiv Detail & Related papers (2024-10-12T07:02:04Z) - Rethinking Out-of-Distribution Detection on Imbalanced Data Distribution [38.844580833635725]
We present a training-time regularization technique to mitigate the bias and boost imbalanced OOD detectors across architecture designs.
Our method translates into consistent improvements on the representative CIFAR10-LT, CIFAR100-LT, and ImageNet-LT benchmarks.
arXiv Detail & Related papers (2024-07-23T12:28:59Z) - Revisiting Confidence Estimation: Towards Reliable Failure Prediction [53.79160907725975]
We find a general, widely existing but actually-neglected phenomenon that most confidence estimation methods are harmful for detecting misclassification errors.
We propose to enlarge the confidence gap by finding flat minima, which yields state-of-the-art failure prediction performance.
arXiv Detail & Related papers (2024-03-05T11:44:14Z) - Model-free Test Time Adaptation for Out-Of-Distribution Detection [62.49795078366206]
We propose a Non-Parametric Test Time textbfAdaptation framework for textbfDistribution textbfDetection (abbr)
abbr utilizes online test samples for model adaptation during testing, enhancing adaptability to changing data distributions.
We demonstrate the effectiveness of abbr through comprehensive experiments on multiple OOD detection benchmarks.
arXiv Detail & Related papers (2023-11-28T02:00:47Z) - Adaptive Uncertainty Estimation via High-Dimensional Testing on Latent
Representations [28.875819909902244]
Uncertainty estimation aims to evaluate the confidence of a trained deep neural network.
Existing uncertainty estimation approaches rely on low-dimensional distributional assumptions.
We propose a new framework using data-adaptive high-dimensional hypothesis testing for uncertainty estimation.
arXiv Detail & Related papers (2023-10-25T12:22:18Z) - Beyond AUROC & co. for evaluating out-of-distribution detection
performance [50.88341818412508]
Given their relevance for safe(r) AI, it is important to examine whether the basis for comparing OOD detection methods is consistent with practical needs.
We propose a new metric - Area Under the Threshold Curve (AUTC), which explicitly penalizes poor separation between ID and OOD samples.
arXiv Detail & Related papers (2023-06-26T12:51:32Z) - iDECODe: In-distribution Equivariance for Conformal Out-of-distribution
Detection [24.518698391381204]
Machine learning methods such as deep neural networks (DNNs) often generate incorrect predictions with high confidence.
We propose the new method iDECODe, leveraging in-distribution equivariance for conformal OOD detection.
We demonstrate the efficacy of iDECODe by experiments on image and audio datasets, obtaining state-of-the-art results.
arXiv Detail & Related papers (2022-01-07T05:21:40Z) - Understanding Failures in Out-of-Distribution Detection with Deep
Generative Models [22.11487118547924]
We prove that no method can guarantee performance beyond random chance without assumptions on which out-distributions are relevant.
We highlight the consequences implied by assuming support overlap between in- and out-distributions.
Our results suggest that estimation error is a more plausible explanation than the misalignment between likelihood-based OOD detection and out-distributions of interest.
arXiv Detail & Related papers (2021-07-14T18:00:11Z) - Provably Robust Detection of Out-of-distribution Data (almost) for free [124.14121487542613]
Deep neural networks are known to produce highly overconfident predictions on out-of-distribution (OOD) data.
In this paper we propose a novel method where from first principles we combine a certifiable OOD detector with a standard classifier into an OOD aware classifier.
In this way we achieve the best of two worlds: certifiably adversarially robust OOD detection, even for OOD samples close to the in-distribution, without loss in prediction accuracy and close to state-of-the-art OOD detection performance for non-manipulated OOD data.
arXiv Detail & Related papers (2021-06-08T11:40:49Z) - Out-of-distribution detection for regression tasks: parameter versus
predictor entropy [2.026281591452464]
It is crucial to detect when an instance lies downright too far from the training samples for the machine learning model to be trusted.
For neural networks, one approach to this task consists of learning a diversity of predictors that all can explain the training data.
We propose a new way of estimating the entropy of a distribution on predictors based on nearest neighbors in function space.
arXiv Detail & Related papers (2020-10-24T21:41:21Z) - Robust Out-of-distribution Detection for Neural Networks [51.19164318924997]
We show that existing detection mechanisms can be extremely brittle when evaluating on in-distribution and OOD inputs.
We propose an effective algorithm called ALOE, which performs robust training by exposing the model to both adversarially crafted inlier and outlier examples.
arXiv Detail & Related papers (2020-03-21T17:46:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.