Improving Object Detection with Selective Self-supervised Self-training
- URL: http://arxiv.org/abs/2007.09162v2
- Date: Fri, 24 Jul 2020 19:34:54 GMT
- Title: Improving Object Detection with Selective Self-supervised Self-training
- Authors: Yandong Li, Di Huang, Danfeng Qin, Liqiang Wang, Boqing Gong
- Abstract summary: We study how to leverage Web images to augment human-curated object detection datasets.
We retrieve Web images by image-to-image search, which incurs less domain shift from the curated data than other search methods.
We propose a novel learning method motivated by two parallel lines of work that explore unlabeled data for image classification.
- Score: 62.792445237541145
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study how to leverage Web images to augment human-curated object detection
datasets. Our approach is two-pronged. On the one hand, we retrieve Web images
by image-to-image search, which incurs less domain shift from the curated data
than other search methods. The Web images are diverse, supplying a wide variety
of object poses, appearances, their interactions with the context, etc. On the
other hand, we propose a novel learning method motivated by two parallel lines
of work that explore unlabeled data for image classification: self-training and
self-supervised learning. They fail to improve object detectors in their
vanilla forms due to the domain gap between the Web images and curated
datasets. To tackle this challenge, we propose a selective net to rectify the
supervision signals in Web images. It not only identifies positive bounding
boxes but also creates a safe zone for mining hard negative boxes. We report
state-of-the-art results on detecting backpacks and chairs from everyday
scenes, along with other challenging object classes.
Related papers
- Data Augmentation for Object Detection via Differentiable Neural
Rendering [71.00447761415388]
It is challenging to train a robust object detector when annotated data is scarce.
Existing approaches to tackle this problem include semi-supervised learning that interpolates labeled data from unlabeled data.
We introduce an offline data augmentation method for object detection, which semantically interpolates the training data with novel views.
arXiv Detail & Related papers (2021-03-04T06:31:06Z) - A Simple and Effective Use of Object-Centric Images for Long-Tailed
Object Detection [56.82077636126353]
We take advantage of object-centric images to improve object detection in scene-centric images.
We present a simple yet surprisingly effective framework to do so.
Our approach can improve the object detection (and instance segmentation) accuracy of rare objects by 50% (and 33%) relatively.
arXiv Detail & Related papers (2021-02-17T17:27:21Z) - Instance Localization for Self-supervised Detection Pretraining [68.24102560821623]
We propose a new self-supervised pretext task, called instance localization.
We show that integration of bounding boxes into pretraining promotes better task alignment and architecture alignment for transfer learning.
Experimental results demonstrate that our approach yields state-of-the-art transfer learning results for object detection.
arXiv Detail & Related papers (2021-02-16T17:58:57Z) - Improved Handling of Motion Blur in Online Object Detection [0.0]
We focus on the details of egomotion induced blur.
We explore five classes of remedies, where each targets different potential causes for the performance gap between sharp and blurred images.
The other four classes of remedies address multi-scale texture, out-of-distribution testing, label generation, and conditioning by blur-type.
arXiv Detail & Related papers (2020-11-29T21:58:26Z) - Co-training for On-board Deep Object Detection [0.0]
Best performing deep vision-based object detectors are trained in a supervised manner by relying on human-labeled bounding boxes.
Co-training is a semi-supervised learning method for self-labeling objects in unlabeled images.
We show how co-training is a paradigm worth to pursue for alleviating object labeling, working both alone and together with task-agnostic domain adaptation.
arXiv Detail & Related papers (2020-08-12T19:08:59Z) - Mining Cross-Image Semantics for Weakly Supervised Semantic Segmentation [128.03739769844736]
Two neural co-attentions are incorporated into the classifier to capture cross-image semantic similarities and differences.
In addition to boosting object pattern learning, the co-attention can leverage context from other related images to improve localization map inference.
Our algorithm sets new state-of-the-arts on all these settings, demonstrating well its efficacy and generalizability.
arXiv Detail & Related papers (2020-07-03T21:53:46Z) - Exploring Bottom-up and Top-down Cues with Attentive Learning for Webly
Supervised Object Detection [76.9756607002489]
We propose a novel webly supervised object detection (WebSOD) method for novel classes.
Our proposed method combines bottom-up and top-down cues for novel class detection.
We demonstrate our proposed method on PASCAL VOC dataset with three different novel/base splits.
arXiv Detail & Related papers (2020-03-22T03:11:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.