OOWL500: Overcoming Dataset Collection Bias in the Wild
- URL: http://arxiv.org/abs/2108.10992v1
- Date: Tue, 24 Aug 2021 23:22:53 GMT
- Title: OOWL500: Overcoming Dataset Collection Bias in the Wild
- Authors: Brandon Leung, Chih-Hui Ho, Amir Persekian, David Orozco, Yen Chang,
Erik Sandstrom, Bo Liu, Nuno Vasconcelos
- Abstract summary: The hypothesis that image datasets gathered online "in the wild" can produce biased object recognizers is studied.
A new "in the lab" data collection infrastructure is proposed consisting of a drone which captures images as it circles around objects.
It's inexpensive and easily replicable nature may also potentially lead to a scalable data collection effort by the vision community.
- Score: 45.494056340200956
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The hypothesis that image datasets gathered online "in the wild" can produce
biased object recognizers, e.g. preferring professional photography or certain
viewing angles, is studied. A new "in the lab" data collection infrastructure
is proposed consisting of a drone which captures images as it circles around
objects. Crucially, the control provided by this setup and the natural camera
shake inherent to flight mitigate many biases. It's inexpensive and easily
replicable nature may also potentially lead to a scalable data collection
effort by the vision community. The procedure's usefulness is demonstrated by
creating a dataset of Objects Obtained With fLight (OOWL). Denoted as OOWL500,
it contains 120,000 images of 500 objects and is the largest "in the lab" image
dataset available when both number of classes and objects per class are
considered. Furthermore, it has enabled several of new insights on object
recognition. First, a novel adversarial attack strategy is proposed, where
image perturbations are defined in terms of semantic properties such as camera
shake and pose. Indeed, experiments have shown that ImageNet has considerable
amounts of pose and professional photography bias. Second, it is used to show
that the augmentation of in the wild datasets, such as ImageNet, with in the
lab data, such as OOWL500, can significantly decrease these biases, leading to
object recognizers of improved generalization. Third, the dataset is used to
study questions on "best procedures" for dataset collection. It is revealed
that data augmentation with synthetic images does not suffice to eliminate in
the wild datasets biases, and that camera shake and pose diversity play a more
important role in object recognition robustness than previously thought.
Related papers
- 360 in the Wild: Dataset for Depth Prediction and View Synthesis [66.58513725342125]
We introduce a large scale 360$circ$ videos dataset in the wild.
This dataset has been carefully scraped from the Internet and has been captured from various locations worldwide.
Each of the 25K images constituting our dataset is provided with its respective camera's pose and depth map.
arXiv Detail & Related papers (2024-06-27T05:26:38Z) - Retrieval Robust to Object Motion Blur [54.34823913494456]
We propose a method for object retrieval in images that are affected by motion blur.
We present the first large-scale datasets for blurred object retrieval.
Our method outperforms state-of-the-art retrieval methods on the new blur-retrieval datasets.
arXiv Detail & Related papers (2024-04-27T23:22:39Z) - Bugs in the Data: How ImageNet Misrepresents Biodiversity [98.98950914663813]
We analyze the 13,450 images from 269 classes that represent wild animals in the ImageNet-1k validation set.
We find that many of the classes are ill-defined or overlapping, and that 12% of the images are incorrectly labeled.
We also find that both the wildlife-related labels and images included in ImageNet-1k present significant geographical and cultural biases.
arXiv Detail & Related papers (2022-08-24T17:55:48Z) - ImageSubject: A Large-scale Dataset for Subject Detection [9.430492045581534]
Main subjects usually exist in the images or videos, as they are the objects that the photographer wants to highlight.
Detecting the main subjects is an important technique to help machines understand the content of images and videos.
We present a new dataset with the goal of training models to understand the layout of the objects then to find the main subjects among them.
arXiv Detail & Related papers (2022-01-09T22:49:59Z) - StereOBJ-1M: Large-scale Stereo Image Dataset for 6D Object Pose
Estimation [43.839322860501596]
We present a large-scale stereo RGB image object pose estimation dataset named the $textbfStereOBJ-1M$ dataset.
The dataset is designed to address challenging cases such as object transparency, translucency, and specular reflection.
We propose a novel method for efficiently annotating pose data in a multi-view fashion that allows data capturing in complex and flexible environments.
arXiv Detail & Related papers (2021-09-21T11:56:38Z) - Object-aware Contrastive Learning for Debiased Scene Representation [74.30741492814327]
We develop a novel object-aware contrastive learning framework that localizes objects in a self-supervised manner.
We also introduce two data augmentations based on ContraCAM, object-aware random crop and background mixup, which reduce contextual and background biases during contrastive self-supervised learning.
arXiv Detail & Related papers (2021-07-30T19:24:07Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - FAIR1M: A Benchmark Dataset for Fine-grained Object Recognition in
High-Resolution Remote Sensing Imagery [21.9319970004788]
We propose a novel benchmark dataset with more than 1 million instances and more than 15,000 images for Fine-grAined object recognItion in high-Resolution remote sensing imagery.
All objects in the FAIR1M dataset are annotated with respect to 5 categories and 37 sub-categories by oriented bounding boxes.
arXiv Detail & Related papers (2021-03-09T17:20:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.