Core Risk Minimization using Salient ImageNet
- URL: http://arxiv.org/abs/2203.15566v1
- Date: Mon, 28 Mar 2022 01:53:34 GMT
- Title: Core Risk Minimization using Salient ImageNet
- Authors: Sahil Singla, Mazda Moayeri, Soheil Feizi
- Abstract summary: We introduce the Salient Imagenet dataset with more than 1 million soft masks localizing core and spurious features for all 1000 Imagenet classes.
Using this dataset, we first evaluate the reliance of several Imagenet pretrained models (42 total) on spurious features.
Next, we introduce a new learning paradigm called Core Risk Minimization (CoRM) whose objective ensures that the model predicts a class using its core features.
- Score: 53.616101711801484
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks can be unreliable in the real world especially when they
heavily use spurious features for their predictions. Recently, Singla & Feizi
(2022) introduced the Salient Imagenet dataset by annotating and localizing
core and spurious features of ~52k samples from 232 classes of Imagenet. While
this dataset is useful for evaluating the reliance of pretrained models on
spurious features, its small size limits its usefulness for training models. In
this work, we first introduce the Salient Imagenet-1M dataset with more than 1
million soft masks localizing core and spurious features for all 1000 Imagenet
classes. Using this dataset, we first evaluate the reliance of several Imagenet
pretrained models (42 total) on spurious features and observe that: (i)
transformers are more sensitive to spurious features compared to Convnets, (ii)
zero-shot CLIP transformers are highly susceptible to spurious features. Next,
we introduce a new learning paradigm called Core Risk Minimization (CoRM) whose
objective ensures that the model predicts a class using its core features. We
evaluate different computational approaches for solving CoRM and achieve
significantly higher (+12%) core accuracy (accuracy when non-core regions
corrupted using noise) with no drop in clean accuracy compared to models
trained via Empirical Risk Minimization.
Related papers
- TinyMIM: An Empirical Study of Distilling MIM Pre-trained Models [31.16595289223858]
Masked image modeling (MIM) performs strongly in pre-training large vision Transformers (ViTs)
However, small models that are critical for real-world applications cannot or only marginally benefit from this pre-training approach.
We explore distillation techniques to transfer the success of large MIM-based pre-trained models to smaller ones.
arXiv Detail & Related papers (2023-01-03T18:59:54Z) - Spurious Features Everywhere -- Large-Scale Detection of Harmful
Spurious Features in ImageNet [36.48282338829549]
In this paper, we develop a framework that allows us to systematically identify spurious features in large datasets like ImageNet.
We validate our results by showing that presence of the harmful spurious feature of a class alone is sufficient to trigger the prediction of that class.
We introduce SpuFix as a simple mitigation method to reduce the dependence of any ImageNet classifier on previously identified harmful spurious features.
arXiv Detail & Related papers (2022-12-09T14:23:25Z) - DeepDC: Deep Distance Correlation as a Perceptual Image Quality
Evaluator [53.57431705309919]
ImageNet pre-trained deep neural networks (DNNs) show notable transferability for building effective image quality assessment (IQA) models.
We develop a novel full-reference IQA (FR-IQA) model based exclusively on pre-trained DNN features.
We conduct comprehensive experiments to demonstrate the superiority of the proposed quality model on five standard IQA datasets.
arXiv Detail & Related papers (2022-11-09T14:57:27Z) - DOMINO: Domain-aware Model Calibration in Medical Image Segmentation [51.346121016559024]
Modern deep neural networks are poorly calibrated, compromising trustworthiness and reliability.
We propose DOMINO, a domain-aware model calibration method that leverages the semantic confusability and hierarchical similarity between class labels.
Our results show that DOMINO-calibrated deep neural networks outperform non-calibrated models and state-of-the-art morphometric methods in head image segmentation.
arXiv Detail & Related papers (2022-09-13T15:31:52Z) - Corrupted Image Modeling for Self-Supervised Visual Pre-Training [103.99311611776697]
We introduce Corrupted Image Modeling (CIM) for self-supervised visual pre-training.
CIM uses an auxiliary generator with a small trainable BEiT to corrupt the input image instead of using artificial mask tokens.
After pre-training, the enhancer can be used as a high-capacity visual encoder for downstream tasks.
arXiv Detail & Related papers (2022-02-07T17:59:04Z) - A Comprehensive Study of Image Classification Model Sensitivity to
Foregrounds, Backgrounds, and Visual Attributes [58.633364000258645]
We call this dataset RIVAL10 consisting of roughly $26k$ instances over $10$ classes.
We evaluate the sensitivity of a broad set of models to noise corruptions in foregrounds, backgrounds and attributes.
In our analysis, we consider diverse state-of-the-art architectures (ResNets, Transformers) and training procedures (CLIP, SimCLR, DeiT, Adversarial Training)
arXiv Detail & Related papers (2022-01-26T06:31:28Z) - Towards Robustness of Neural Networks [0.0]
We introduce ImageNet-A/O and ImageNet-R as well as a synthetic environment and testing suite we called CAOS.
All of the datasets were created for testing robustness and measuring progress in robustness.
We build off of simple baselines in the form of Maximum Logit, and Typicality Score as well as create a novel data augmentation method in the form of DeepAugment.
arXiv Detail & Related papers (2021-12-30T19:41:10Z) - Shape-Texture Debiased Neural Network Training [50.6178024087048]
Convolutional Neural Networks are often biased towards either texture or shape, depending on the training dataset.
We develop an algorithm for shape-texture debiased learning.
Experiments show that our method successfully improves model performance on several image recognition benchmarks.
arXiv Detail & Related papers (2020-10-12T19:16:12Z) - Impact of ImageNet Model Selection on Domain Adaptation [26.016647703500883]
We investigate how different ImageNet models affect transfer accuracy on domain adaptation problems.
A higher accuracy ImageNet model produces better features, and leads to higher accuracy on domain adaptation problems.
We also examine the architecture of each neural network to find the best layer for feature extraction.
arXiv Detail & Related papers (2020-02-06T23:58:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.