A framework for benchmarking class-out-of-distribution detection and its
application to ImageNet
- URL: http://arxiv.org/abs/2302.11893v1
- Date: Thu, 23 Feb 2023 09:57:48 GMT
- Title: A framework for benchmarking class-out-of-distribution detection and its
application to ImageNet
- Authors: Ido Galil, Mohammed Dabbah, Ran El-Yaniv
- Abstract summary: We present a novel framework to benchmark the ability of image classifiers to detect classout-of-distribution instances.
We apply this technique to ImageNet, and 525 pretrained, publicly available, ImageNet-1k classifiers.
- Score: 15.929238800072195
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: When deployed for risk-sensitive tasks, deep neural networks must be able to
detect instances with labels from outside the distribution for which they were
trained. In this paper we present a novel framework to benchmark the ability of
image classifiers to detect class-out-of-distribution instances (i.e.,
instances whose true labels do not appear in the training distribution) at
various levels of detection difficulty. We apply this technique to ImageNet,
and benchmark 525 pretrained, publicly available, ImageNet-1k classifiers. The
code for generating a benchmark for any ImageNet-1k classifier, along with the
benchmarks prepared for the above-mentioned 525 models is available at
https://github.com/mdabbah/COOD_benchmarking.
The usefulness of the proposed framework and its advantage over alternative
existing benchmarks is demonstrated by analyzing the results obtained for these
models, which reveals numerous novel observations including: (1) knowledge
distillation consistently improves class-out-of-distribution (C-OOD) detection
performance; (2) a subset of ViTs performs better C-OOD detection than any
other model; (3) the language--vision CLIP model achieves good zero-shot
detection performance, with its best instance outperforming 96% of all other
models evaluated; (4) accuracy and in-distribution ranking are positively
correlated to C-OOD detection; and (5) we compare various confidence functions
for C-OOD detection. Our companion paper, also published in ICLR 2023 (What Can
We Learn From The Selective Prediction And Uncertainty Estimation Performance
Of 523 Imagenet Classifiers), examines the uncertainty estimation performance
(ranking, calibration, and selective prediction performance) of these
classifiers in an in-distribution setting.
Related papers
- Credal Wrapper of Model Averaging for Uncertainty Estimation on Out-Of-Distribution Detection [5.19656787424626]
This paper presents an innovative approach, called credal wrapper, to formulating a credal set representation of model averaging for Bayesian neural networks (BNNs) and deep ensembles.
Given a finite collection of single distributions derived from BNNs or deep ensembles, the proposed approach extracts an upper and a lower probability bound per class.
Compared to BNN and deep ensemble baselines, the proposed credal representation methodology exhibits superior performance in uncertainty estimation.
arXiv Detail & Related papers (2024-05-23T20:51:22Z) - Investigating the Limitation of CLIP Models: The Worst-Performing
Categories [53.360239882501325]
Contrastive Language-Image Pre-training (CLIP) provides a foundation model by integrating natural language into visual concepts.
It is usually expected that satisfactory overall accuracy can be achieved across numerous domains through well-designed textual prompts.
However, we found that their performance in the worst categories is significantly inferior to the overall performance.
arXiv Detail & Related papers (2023-10-05T05:37:33Z) - Nearest Neighbor Guidance for Out-of-Distribution Detection [18.851275688720108]
We propose Nearest Neighbor Guidance (NNGuide) for detecting out-of-distribution (OOD) samples.
NNGuide reduces the overconfidence of OOD samples while preserving the fine-grained capability of the classifier-based score.
Our results demonstrate that NNGuide provides a significant performance improvement on the base detection scores.
arXiv Detail & Related papers (2023-09-26T12:40:35Z) - Adapting Contrastive Language-Image Pretrained (CLIP) Models for
Out-of-Distribution Detection [1.597617022056624]
We present a comprehensive experimental study on pretrained feature extractors for visual out-of-distribution (OOD) detection.
We propose a new simple and scalable method called textitpseudo-label probing (PLP) that adapts vision-language models for OOD detection.
arXiv Detail & Related papers (2023-03-10T10:02:18Z) - What Can We Learn From The Selective Prediction And Uncertainty
Estimation Performance Of 523 Imagenet Classifiers [15.929238800072195]
We present a novel study of selective prediction and the uncertainty estimation performance of 523 existing pretrained deep ImageNet classifiers.
We find that distillation-based training regimes consistently yield better uncertainty estimations than other training schemes.
For example, we discovered an unprecedented 99% top-1 selective accuracy on ImageNet at 47% coverage.
arXiv Detail & Related papers (2023-02-23T09:25:28Z) - Effective Robustness against Natural Distribution Shifts for Models with
Different Training Data [113.21868839569]
"Effective robustness" measures the extra out-of-distribution robustness beyond what can be predicted from the in-distribution (ID) performance.
We propose a new evaluation metric to evaluate and compare the effective robustness of models trained on different data.
arXiv Detail & Related papers (2023-02-02T19:28:41Z) - Boosting Out-of-Distribution Detection with Multiple Pre-trained Models [41.66566916581451]
Post hoc detection utilizing pre-trained models has shown promising performance and can be scaled to large-scale problems.
We propose a detection enhancement method by ensembling multiple detection decisions derived from a zoo of pre-trained models.
Our method substantially improves the relative performance by 65.40% and 26.96% on the CIFAR10 and ImageNet benchmarks.
arXiv Detail & Related papers (2022-12-24T12:11:38Z) - mc-BEiT: Multi-choice Discretization for Image BERT Pre-training [52.04866462439979]
Image BERT pre-training with masked image modeling (MIM) is a popular practice to cope with self-supervised representation learning.
We introduce an improved BERT-style image pre-training method, namely mc-BEiT, which performs MIM proxy tasks towards eased and refined multi-choice training objectives.
arXiv Detail & Related papers (2022-03-29T09:08:18Z) - Measuring Self-Supervised Representation Quality for Downstream
Classification using Discriminative Features [56.89813105411331]
We study the representation space of state-of-the-art self-supervised models including SimCLR, SwaV, MoCo, BYOL, DINO, SimSiam, VICReg and Barlow Twins.
We propose Self-Supervised Representation Quality Score (or Q-Score), an unsupervised score that can reliably predict if a given sample is likely to be mis-classified.
Fine-tuning with Q-Score regularization can boost the linear probing accuracy of SSL models by up to 5.8% on ImageNet-100 and 3.7% on ImageNet-1K.
arXiv Detail & Related papers (2022-03-03T17:48:23Z) - Generalized Focal Loss: Learning Qualified and Distributed Bounding
Boxes for Dense Object Detection [85.53263670166304]
One-stage detector basically formulates object detection as dense classification and localization.
Recent trend for one-stage detectors is to introduce an individual prediction branch to estimate the quality of localization.
This paper delves into the representations of the above three fundamental elements: quality estimation, classification and localization.
arXiv Detail & Related papers (2020-06-08T07:24:33Z) - I Am Going MAD: Maximum Discrepancy Competition for Comparing
Classifiers Adaptively [135.7695909882746]
We name the MAximum Discrepancy (MAD) competition.
We adaptively sample a small test set from an arbitrarily large corpus of unlabeled images.
Human labeling on the resulting model-dependent image sets reveals the relative performance of the competing classifiers.
arXiv Detail & Related papers (2020-02-25T03:32:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.