What Can We Learn From The Selective Prediction And Uncertainty
Estimation Performance Of 523 Imagenet Classifiers
- URL: http://arxiv.org/abs/2302.11874v1
- Date: Thu, 23 Feb 2023 09:25:28 GMT
- Title: What Can We Learn From The Selective Prediction And Uncertainty
Estimation Performance Of 523 Imagenet Classifiers
- Authors: Ido Galil, Mohammed Dabbah, Ran El-Yaniv
- Abstract summary: We present a novel study of selective prediction and the uncertainty estimation performance of 523 existing pretrained deep ImageNet classifiers.
We find that distillation-based training regimes consistently yield better uncertainty estimations than other training schemes.
For example, we discovered an unprecedented 99% top-1 selective accuracy on ImageNet at 47% coverage.
- Score: 15.929238800072195
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: When deployed for risk-sensitive tasks, deep neural networks must include an
uncertainty estimation mechanism. Here we examine the relationship between deep
architectures and their respective training regimes, with their corresponding
selective prediction and uncertainty estimation performance. We consider some
of the most popular estimation performance metrics previously proposed
including AUROC, ECE, AURC as well as coverage for selective accuracy
constraint. We present a novel and comprehensive study of selective prediction
and the uncertainty estimation performance of 523 existing pretrained deep
ImageNet classifiers that are available in popular repositories. We identify
numerous and previously unknown factors that affect uncertainty estimation and
examine the relationships between the different metrics. We find that
distillation-based training regimes consistently yield better uncertainty
estimations than other training schemes such as vanilla training, pretraining
on a larger dataset and adversarial training. Moreover, we find a subset of ViT
models that outperform any other models in terms of uncertainty estimation
performance. For example, we discovered an unprecedented 99% top-1 selective
accuracy on ImageNet at 47% coverage (and 95% top-1 accuracy at 80%) for a ViT
model, whereas a competing EfficientNet-V2-XL cannot obtain these accuracy
constraints at any level of coverage. Our companion paper, also published in
ICLR 2023 (A framework for benchmarking class-out-of-distribution detection and
its application to ImageNet), examines the performance of these classifiers in
a class-out-of-distribution setting.
Related papers
- Leveraging Uncertainty Estimates To Improve Classifier Performance [4.4951754159063295]
Binary classification involves predicting the label of an instance based on whether the model score for the positive class exceeds a threshold chosen based on the application requirements.
However, model scores are often not aligned with the true positivity rate.
This is especially true when the training involves a differential sampling across classes or there is distributional drift between train and test settings.
arXiv Detail & Related papers (2023-11-20T12:40:25Z) - Learning Sample Difficulty from Pre-trained Models for Reliable
Prediction [55.77136037458667]
We propose to utilize large-scale pre-trained models to guide downstream model training with sample difficulty-aware entropy regularization.
We simultaneously improve accuracy and uncertainty calibration across challenging benchmarks.
arXiv Detail & Related papers (2023-04-20T07:29:23Z) - A framework for benchmarking class-out-of-distribution detection and its
application to ImageNet [15.929238800072195]
We present a novel framework to benchmark the ability of image classifiers to detect classout-of-distribution instances.
We apply this technique to ImageNet, and 525 pretrained, publicly available, ImageNet-1k classifiers.
arXiv Detail & Related papers (2023-02-23T09:57:48Z) - Post-hoc Uncertainty Learning using a Dirichlet Meta-Model [28.522673618527417]
We propose a novel Bayesian meta-model to augment pre-trained models with better uncertainty quantification abilities.
Our proposed method requires no additional training data and is flexible enough to quantify different uncertainties.
We demonstrate our proposed meta-model approach's flexibility and superior empirical performance on these applications.
arXiv Detail & Related papers (2022-12-14T17:34:11Z) - Estimating Model Performance under Domain Shifts with Class-Specific
Confidence Scores [25.162667593654206]
We introduce class-wise calibration within the framework of performance estimation for imbalanced datasets.
We conduct experiments on four tasks and find the proposed modifications consistently improve the estimation accuracy for imbalanced datasets.
arXiv Detail & Related papers (2022-07-20T15:04:32Z) - BayesCap: Bayesian Identity Cap for Calibrated Uncertainty in Frozen
Neural Networks [50.15201777970128]
We propose BayesCap that learns a Bayesian identity mapping for the frozen model, allowing uncertainty estimation.
BayesCap is a memory-efficient method that can be trained on a small fraction of the original dataset.
We show the efficacy of our method on a wide variety of tasks with a diverse set of architectures.
arXiv Detail & Related papers (2022-07-14T12:50:09Z) - Which models are innately best at uncertainty estimation? [15.929238800072195]
Deep neural networks must be equipped with an uncertainty estimation mechanism when deployed for risk-sensitive tasks.
This paper studies the relationship between deep architectures and their training regimes with their corresponding selective prediction and uncertainty estimation performance.
arXiv Detail & Related papers (2022-06-05T11:15:35Z) - NUQ: Nonparametric Uncertainty Quantification for Deterministic Neural
Networks [151.03112356092575]
We show the principled way to measure the uncertainty of predictions for a classifier based on Nadaraya-Watson's nonparametric estimate of the conditional label distribution.
We demonstrate the strong performance of the method in uncertainty estimation tasks on a variety of real-world image datasets.
arXiv Detail & Related papers (2022-02-07T12:30:45Z) - Dense Uncertainty Estimation [62.23555922631451]
In this paper, we investigate neural networks and uncertainty estimation techniques to achieve both accurate deterministic prediction and reliable uncertainty estimation.
We work on two types of uncertainty estimations solutions, namely ensemble based methods and generative model based methods, and explain their pros and cons while using them in fully/semi/weakly-supervised framework.
arXiv Detail & Related papers (2021-10-13T01:23:48Z) - Learning Accurate Dense Correspondences and When to Trust Them [161.76275845530964]
We aim to estimate a dense flow field relating two images, coupled with a robust pixel-wise confidence map.
We develop a flexible probabilistic approach that jointly learns the flow prediction and its uncertainty.
Our approach obtains state-of-the-art results on challenging geometric matching and optical flow datasets.
arXiv Detail & Related papers (2021-01-05T18:54:11Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.