From ImageNet to Image Classification: Contextualizing Progress on
Benchmarks
- URL: http://arxiv.org/abs/2005.11295v1
- Date: Fri, 22 May 2020 17:39:16 GMT
- Title: From ImageNet to Image Classification: Contextualizing Progress on
Benchmarks
- Authors: Dimitris Tsipras, Shibani Santurkar, Logan Engstrom, Andrew Ilyas,
Aleksander Madry
- Abstract summary: We study how specific design choices in the ImageNet creation process impact the fidelity of the resulting dataset.
Our analysis pinpoints how a noisy data collection pipeline can lead to a systematic misalignment between the resulting benchmark and the real-world task it serves as a proxy for.
- Score: 99.19183528305598
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Building rich machine learning datasets in a scalable manner often
necessitates a crowd-sourced data collection pipeline. In this work, we use
human studies to investigate the consequences of employing such a pipeline,
focusing on the popular ImageNet dataset. We study how specific design choices
in the ImageNet creation process impact the fidelity of the resulting
dataset---including the introduction of biases that state-of-the-art models
exploit. Our analysis pinpoints how a noisy data collection pipeline can lead
to a systematic misalignment between the resulting benchmark and the real-world
task it serves as a proxy for. Finally, our findings emphasize the need to
augment our current model training and evaluation toolkit to take such
misalignments into account. To facilitate further research, we release our
refined ImageNet annotations at https://github.com/MadryLab/ImageNetMultiLabel.
Related papers
- Rethinking Image Super-Resolution from Training Data Perspectives [54.28824316574355]
We investigate the understudied effect of the training data used for image super-resolution (SR)
With this, we propose an automated image evaluation pipeline.
We find that datasets with (i) low compression artifacts, (ii) high within-image diversity as judged by the number of different objects, and (iii) a large number of images from ImageNet or PASS all positively affect SR performance.
arXiv Detail & Related papers (2024-09-01T16:25:04Z) - Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - ContextMix: A context-aware data augmentation method for industrial
visual inspection systems [21.35653674563189]
We introduce ContextMix, a method tailored for industrial applications and benchmark datasets.
ContextMix generates novel data by resizing entire images and integrating them into other images within the batch.
We evaluate its effectiveness across classification, detection, and segmentation tasks using various network architectures on public benchmark datasets.
arXiv Detail & Related papers (2024-01-18T15:15:32Z) - DreamTeacher: Pretraining Image Backbones with Deep Generative Models [103.62397699392346]
We introduce a self-supervised feature representation learning framework that utilizes generative networks for pre-training downstream image backbones.
We investigate two types of knowledge distillation: 1) distilling learned generative features onto target image backbones as an alternative to pretraining these backbones on large labeled datasets such as ImageNet.
We empirically find that our DreamTeacher significantly outperforms existing self-supervised representation learning approaches across the board.
arXiv Detail & Related papers (2023-07-14T17:17:17Z) - BigDatasetGAN: Synthesizing ImageNet with Pixel-wise Annotations [89.42397034542189]
We synthesize a large labeled dataset via a generative adversarial network (GAN)
We take image samples from the class-conditional generative model BigGAN trained on ImageNet, and manually annotate 5 images per class, for all 1k classes.
We create a new ImageNet benchmark by labeling an additional set of 8k real images and evaluate segmentation performance in a variety of settings.
arXiv Detail & Related papers (2022-01-12T20:28:34Z) - Learning Co-segmentation by Segment Swapping for Retrieval and Discovery [67.6609943904996]
The goal of this work is to efficiently identify visually similar patterns from a pair of images.
We generate synthetic training pairs by selecting object segments in an image and copy-pasting them into another image.
We show our approach provides clear improvements for artwork details retrieval on the Brueghel dataset.
arXiv Detail & Related papers (2021-10-29T16:51:16Z) - Exploiting the relationship between visual and textual features in
social networks for image classification with zero-shot deep learning [0.0]
In this work, we propose a classifier ensemble based on the transferable learning capabilities of the CLIP neural network architecture.
Our experiments, based on image classification tasks according to the labels of the Places dataset, are performed by first considering only the visual part.
Considering the associated texts to the images can help to improve the accuracy depending on the goal.
arXiv Detail & Related papers (2021-07-08T10:54:59Z) - Automated Cleanup of the ImageNet Dataset by Model Consensus,
Explainability and Confident Learning [0.0]
ImageNet was the backbone of various convolutional neural networks (CNNs) trained on ILSVRC12Net.
This paper describes automated applications based on model consensus, explainability and confident learning to correct labeling mistakes.
The ImageNet-Clean improves the model performance by 2-2.4 % for SqueezeNet and EfficientNet-B0 models.
arXiv Detail & Related papers (2021-03-30T13:16:35Z) - Unsupervised machine learning via transfer learning and k-means
clustering to classify materials image data [0.0]
This paper demonstrates how to construct, use, and evaluate a high performance unsupervised machine learning system for classifying images.
We use the VGG16 convolutional neural network pre-trained on the ImageNet dataset of natural images to extract feature representations for each micrograph.
The approach achieves $99.4% pm 0.16%$ accuracy, and the resulting model can be used to classify new images without retraining.
arXiv Detail & Related papers (2020-07-16T14:36:04Z) - Multi-task pre-training of deep neural networks for digital pathology [8.74883469030132]
We first assemble and transform many digital pathology datasets into a pool of 22 classification tasks and almost 900k images.
We show that our models used as feature extractors either improve significantly over ImageNet pre-trained models or provide comparable performance.
arXiv Detail & Related papers (2020-05-05T08:50:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.