Raising the Bar on the Evaluation of Out-of-Distribution Detection
- URL: http://arxiv.org/abs/2209.11960v1
- Date: Sat, 24 Sep 2022 08:48:36 GMT
- Title: Raising the Bar on the Evaluation of Out-of-Distribution Detection
- Authors: Jishnu Mukhoti, Tsung-Yu Lin, Bor-Chun Chen, Ashish Shah, Philip H.S.
Torr, Puneet K. Dokania, Ser-Nam Lim
- Abstract summary: We define 2 categories of OoD data using the subtly different concepts of perceptual/visual and semantic similarity to in-distribution (iD) data.
We propose a GAN based framework for generating OoD samples from each of these 2 categories, given an iD dataset.
We show that a) state-of-the-art OoD detection methods which perform exceedingly well on conventional benchmarks are significantly less robust to our proposed benchmark.
- Score: 88.70479625837152
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In image classification, a lot of development has happened in detecting
out-of-distribution (OoD) data. However, most OoD detection methods are
evaluated on a standard set of datasets, arbitrarily different from training
data. There is no clear definition of what forms a ``good" OoD dataset.
Furthermore, the state-of-the-art OoD detection methods already achieve near
perfect results on these standard benchmarks. In this paper, we define 2
categories of OoD data using the subtly different concepts of perceptual/visual
and semantic similarity to in-distribution (iD) data. We define Near OoD
samples as perceptually similar but semantically different from iD samples, and
Shifted samples as points which are visually different but semantically akin to
iD data. We then propose a GAN based framework for generating OoD samples from
each of these 2 categories, given an iD dataset. Through extensive experiments
on MNIST, CIFAR-10/100 and ImageNet, we show that a) state-of-the-art OoD
detection methods which perform exceedingly well on conventional benchmarks are
significantly less robust to our proposed benchmark. Moreover, b) models
performing well on our setup also perform well on conventional real-world OoD
detection benchmarks and vice versa, thereby indicating that one might not even
need a separate OoD set, to reliably evaluate performance in OoD detection.
Related papers
- Revisiting Deep Ensemble for Out-of-Distribution Detection: A Loss Landscape Perspective [28.545919238058982]
Out-of-Distribution (OoD) detection methods address to detect OoD samples from In-Distribution (InD) data.
We propose a new perspective upon loss landscape and mode ensemble to investigate OoD detection.
arXiv Detail & Related papers (2023-10-22T08:11:51Z) - SEE-OoD: Supervised Exploration For Enhanced Out-of-Distribution
Detection [11.05254400092658]
We propose a Wasserstein-score-based generative adversarial training scheme to enhance OoD detection accuracy.
Specifically, the generator explores OoD spaces and generates synthetic OoD samples using feedback from the discriminator.
We demonstrate that the proposed method outperforms state-of-the-art techniques on various computer vision datasets.
arXiv Detail & Related papers (2023-10-12T05:20:18Z) - Interpretable Out-Of-Distribution Detection Using Pattern Identification [0.0]
Out-of-distribution (OoD) detection for data-based programs is a goal of paramount importance.
Common approaches in the literature tend to train detectors requiring inside-of-distribution (in-distribution, or IoD) and OoD validation samples.
We propose to use existing work from the field of explainable AI, namely the PARTICUL pattern identification algorithm, in order to build more interpretable and robust OoD detectors.
arXiv Detail & Related papers (2023-01-24T15:35:54Z) - GOOD-D: On Unsupervised Graph Out-Of-Distribution Detection [67.90365841083951]
We develop a new graph contrastive learning framework GOOD-D for detecting OOD graphs without using any ground-truth labels.
GOOD-D is able to capture the latent ID patterns and accurately detect OOD graphs based on the semantic inconsistency in different granularities.
As a pioneering work in unsupervised graph-level OOD detection, we build a comprehensive benchmark to compare our proposed approach with different state-of-the-art methods.
arXiv Detail & Related papers (2022-11-08T12:41:58Z) - Benchmarking Deep Models for Salient Object Detection [67.07247772280212]
We construct a general SALient Object Detection (SALOD) benchmark to conduct a comprehensive comparison among several representative SOD methods.
In the above experiments, we find that existing loss functions usually specialized in some metrics but reported inferior results on the others.
We propose a novel Edge-Aware (EA) loss that promotes deep networks to learn more discriminative features by integrating both pixel- and image-level supervision signals.
arXiv Detail & Related papers (2022-02-07T03:43:16Z) - DOI: Divergence-based Out-of-Distribution Indicators via Deep Generative
Models [6.617664042202313]
OoD (out-of-distribution) indicators based on deep generative models are proposed recently and are shown to work well on small datasets.
We conduct the first large collection of benchmarks (containing 92 dataset pairs) for existing OoD indicators and observe that none perform well.
We propose a novel theoretical framework, DOI, for divergence-based Out-of-Distribution indicators (instead of traditional likelihood-based) in deep generative models.
arXiv Detail & Related papers (2021-08-12T02:49:54Z) - Entropy Maximization and Meta Classification for Out-Of-Distribution
Detection in Semantic Segmentation [7.305019142196585]
"Out-of-distribution" (OoD) samples are crucial for many applications such as automated driving.
A natural baseline approach to OoD detection is to threshold on the pixel-wise softmax entropy.
We present a two-step procedure that significantly improves that approach.
arXiv Detail & Related papers (2020-12-09T11:01:06Z) - NADS: Neural Architecture Distribution Search for Uncertainty Awareness [79.18710225716791]
Machine learning (ML) systems often encounter Out-of-Distribution (OoD) errors when dealing with testing data coming from a distribution different from training data.
Existing OoD detection approaches are prone to errors and even sometimes assign higher likelihoods to OoD samples.
We propose Neural Architecture Distribution Search (NADS) to identify common building blocks among all uncertainty-aware architectures.
arXiv Detail & Related papers (2020-06-11T17:39:07Z) - Generalized ODIN: Detecting Out-of-distribution Image without Learning
from Out-of-distribution Data [87.61504710345528]
We propose two strategies for freeing a neural network from tuning with OoD data, while improving its OoD detection performance.
We specifically propose to decompose confidence scoring as well as a modified input pre-processing method.
Our further analysis on a larger scale image dataset shows that the two types of distribution shifts, specifically semantic shift and non-semantic shift, present a significant difference.
arXiv Detail & Related papers (2020-02-26T04:18:25Z) - Stance Detection Benchmark: How Robust Is Your Stance Detection? [65.91772010586605]
Stance Detection (StD) aims to detect an author's stance towards a certain topic or claim.
We introduce a StD benchmark that learns from ten StD datasets of various domains in a multi-dataset learning setting.
Within this benchmark setup, we are able to present new state-of-the-art results on five of the datasets.
arXiv Detail & Related papers (2020-01-06T13:37:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.