Co-training for On-board Deep Object Detection
- URL: http://arxiv.org/abs/2008.05534v1
- Date: Wed, 12 Aug 2020 19:08:59 GMT
- Title: Co-training for On-board Deep Object Detection
- Authors: Gabriel Villalonga and Antonio M. Lopez
- Abstract summary: Best performing deep vision-based object detectors are trained in a supervised manner by relying on human-labeled bounding boxes.
Co-training is a semi-supervised learning method for self-labeling objects in unlabeled images.
We show how co-training is a paradigm worth to pursue for alleviating object labeling, working both alone and together with task-agnostic domain adaptation.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Providing ground truth supervision to train visual models has been a
bottleneck over the years, exacerbated by domain shifts which degenerate the
performance of such models. This was the case when visual tasks relied on
handcrafted features and shallow machine learning and, despite its
unprecedented performance gains, the problem remains open within the deep
learning paradigm due to its data-hungry nature. Best performing deep
vision-based object detectors are trained in a supervised manner by relying on
human-labeled bounding boxes which localize class instances (i.e.objects)
within the training images.Thus, object detection is one of such tasks for
which human labeling is a major bottleneck. In this paper, we assess
co-training as a semi-supervised learning method for self-labeling objects in
unlabeled images, so reducing the human-labeling effort for developing deep
object detectors. Our study pays special attention to a scenario involving
domain shift; in particular, when we have automatically generated virtual-world
images with object bounding boxes and we have real-world images which are
unlabeled. Moreover, we are particularly interested in using co-training for
deep object detection in the context of driver assistance systems and/or
self-driving vehicles. Thus, using well-established datasets and protocols for
object detection in these application contexts, we will show how co-training is
a paradigm worth to pursue for alleviating object labeling, working both alone
and together with task-agnostic domain adaptation.
Related papers
- PEEKABOO: Hiding parts of an image for unsupervised object localization [7.161489957025654]
Localizing objects in an unsupervised manner poses significant challenges due to the absence of key visual information.
We propose a single-stage learning framework, dubbed PEEKABOO, for unsupervised object localization.
The key idea is to selectively hide parts of an image and leverage the remaining image information to infer the location of objects without explicit supervision.
arXiv Detail & Related papers (2024-07-24T20:35:20Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Self-Supervised Pretraining on Satellite Imagery: a Case Study on
Label-Efficient Vehicle Detection [0.0]
We study in-domain self-supervised representation learning for object detection on very high resolution optical satellite imagery.
We use the large land use classification dataset Functional Map of the World to pretrain representations with an extension of the Momentum Contrast framework.
We then investigate this model's transferability on a real-world task of fine-grained vehicle detection and classification on Preligens proprietary data.
arXiv Detail & Related papers (2022-10-21T08:41:22Z) - Unsupervised Domain Adaption of Object Detectors: A Survey [87.08473838767235]
Recent advances in deep learning have led to the development of accurate and efficient models for various computer vision applications.
Learning highly accurate models relies on the availability of datasets with a large number of annotated images.
Due to this, model performance drops drastically when evaluated on label-scarce datasets having visually distinct images.
arXiv Detail & Related papers (2021-05-27T23:34:06Z) - A Simple and Effective Use of Object-Centric Images for Long-Tailed
Object Detection [56.82077636126353]
We take advantage of object-centric images to improve object detection in scene-centric images.
We present a simple yet surprisingly effective framework to do so.
Our approach can improve the object detection (and instance segmentation) accuracy of rare objects by 50% (and 33%) relatively.
arXiv Detail & Related papers (2021-02-17T17:27:21Z) - Instance Localization for Self-supervised Detection Pretraining [68.24102560821623]
We propose a new self-supervised pretext task, called instance localization.
We show that integration of bounding boxes into pretraining promotes better task alignment and architecture alignment for transfer learning.
Experimental results demonstrate that our approach yields state-of-the-art transfer learning results for object detection.
arXiv Detail & Related papers (2021-02-16T17:58:57Z) - Self-supervised Segmentation via Background Inpainting [96.10971980098196]
We introduce a self-supervised detection and segmentation approach that can work with single images captured by a potentially moving camera.
We exploit a self-supervised loss function that we exploit to train a proposal-based segmentation network.
We apply our method to human detection and segmentation in images that visually depart from those of standard benchmarks and outperform existing self-supervised methods.
arXiv Detail & Related papers (2020-11-11T08:34:40Z) - Improving Object Detection with Selective Self-supervised Self-training [62.792445237541145]
We study how to leverage Web images to augment human-curated object detection datasets.
We retrieve Web images by image-to-image search, which incurs less domain shift from the curated data than other search methods.
We propose a novel learning method motivated by two parallel lines of work that explore unlabeled data for image classification.
arXiv Detail & Related papers (2020-07-17T18:05:01Z) - A Deep Learning Approach to Object Affordance Segmentation [31.221897360610114]
We design an autoencoder that infers pixel-wise affordance labels in both videos and static images.
Our model surpasses the need for object labels and bounding boxes by using a soft-attention mechanism.
We show that our model achieves competitive results compared to strongly supervised methods on SOR3D-AFF.
arXiv Detail & Related papers (2020-04-18T15:34:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.