A Simple and Effective Use of Object-Centric Images for Long-Tailed
Object Detection
- URL: http://arxiv.org/abs/2102.08884v1
- Date: Wed, 17 Feb 2021 17:27:21 GMT
- Title: A Simple and Effective Use of Object-Centric Images for Long-Tailed
Object Detection
- Authors: Cheng Zhang, Tai-Yu Pan, Yandong Li, Hexiang Hu, Dong Xuan, Soravit
Changpinyo, Boqing Gong, Wei-Lun Chao
- Abstract summary: We take advantage of object-centric images to improve object detection in scene-centric images.
We present a simple yet surprisingly effective framework to do so.
Our approach can improve the object detection (and instance segmentation) accuracy of rare objects by 50% (and 33%) relatively.
- Score: 56.82077636126353
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object frequencies in daily scenes follow a long-tailed distribution. Many
objects do not appear frequently enough in scene-centric images (e.g.,
sightseeing, street views) for us to train accurate object detectors. In
contrast, these objects are captured at a higher frequency in object-centric
images, which are intended to picture the objects of interest. Motivated by
this phenomenon, we propose to take advantage of the object-centric images to
improve object detection in scene-centric images. We present a simple yet
surprisingly effective framework to do so. On the one hand, our approach turns
an object-centric image into a useful training example for object detection in
scene-centric images by mitigating the domain gap between the two image sources
in both the input and label space. On the other hand, our approach employs a
multi-stage procedure to train the object detector, such that the detector
learns the diverse object appearances from object-centric images while being
tied to the application domain of scene-centric images. On the LVIS dataset,
our approach can improve the object detection (and instance segmentation)
accuracy of rare objects by 50% (and 33%) relatively, without sacrificing the
performance of other classes.
Related papers
- Improving Object Detection via Local-global Contrastive Learning [27.660633883387753]
We present a novel image-to-image translation method that specifically targets cross-domain object detection.
We learn to represent objects by contrasting local-global information.
This affords investigation of an under-explored challenge: obtaining performant detection, under domain shifts.
arXiv Detail & Related papers (2024-10-07T14:18:32Z) - Retrieval Robust to Object Motion Blur [54.34823913494456]
We propose a method for object retrieval in images that are affected by motion blur.
We present the first large-scale datasets for blurred object retrieval.
Our method outperforms state-of-the-art retrieval methods on the new blur-retrieval datasets.
arXiv Detail & Related papers (2024-04-27T23:22:39Z) - LOCATE: Localize and Transfer Object Parts for Weakly Supervised
Affordance Grounding [43.157518990171674]
Humans excel at acquiring knowledge through observation.
A key step to acquire this skill is to identify what part of the object affords each action, which is called affordance grounding.
We propose a framework called LOCATE that can identify matching object parts across images.
arXiv Detail & Related papers (2023-03-16T21:47:49Z) - Image Segmentation-based Unsupervised Multiple Objects Discovery [1.7674345486888503]
Unsupervised object discovery aims to localize objects in images.
We propose a fully unsupervised, bottom-up approach, for multiple objects discovery.
We provide state-of-the-art results for both unsupervised class-agnostic object detection and unsupervised image segmentation.
arXiv Detail & Related papers (2022-12-20T09:48:24Z) - Learning to Detect Every Thing in an Open World [139.78830329914135]
We propose a simple yet surprisingly powerful data augmentation and training scheme we call Learning to Detect Every Thing (LDET)
To avoid suppressing hidden objects, background objects that are visible but unlabeled, we paste annotated objects on a background image sampled from a small region of the original image.
LDET leads to significant improvements on many datasets in the open world instance segmentation task.
arXiv Detail & Related papers (2021-12-03T03:56:06Z) - Object-aware Contrastive Learning for Debiased Scene Representation [74.30741492814327]
We develop a novel object-aware contrastive learning framework that localizes objects in a self-supervised manner.
We also introduce two data augmentations based on ContraCAM, object-aware random crop and background mixup, which reduce contextual and background biases during contrastive self-supervised learning.
arXiv Detail & Related papers (2021-07-30T19:24:07Z) - Learning Object Detection from Captions via Textual Scene Attributes [70.90708863394902]
We argue that captions contain much richer information about the image, including attributes of objects and their relations.
We present a method that uses the attributes in this "textual scene graph" to train object detectors.
We empirically demonstrate that the resulting model achieves state-of-the-art results on several challenging object detection datasets.
arXiv Detail & Related papers (2020-09-30T10:59:20Z) - Improving Object Detection with Selective Self-supervised Self-training [62.792445237541145]
We study how to leverage Web images to augment human-curated object detection datasets.
We retrieve Web images by image-to-image search, which incurs less domain shift from the curated data than other search methods.
We propose a novel learning method motivated by two parallel lines of work that explore unlabeled data for image classification.
arXiv Detail & Related papers (2020-07-17T18:05:01Z) - Evaluating Salient Object Detection in Natural Images with Multiple
Objects having Multi-level Saliency [3.464871689508835]
Salient object detection is evaluated using binary ground truth with the labels being salient object class and background.
Our dataset, named SalMoN (saliency in multi-object natural images), has 588 images containing multiple objects.
arXiv Detail & Related papers (2020-03-19T00:06:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.