An Empirical Study of Pseudo-Labeling for Image-based 3D Object
Detection
- URL: http://arxiv.org/abs/2208.07137v1
- Date: Mon, 15 Aug 2022 12:17:46 GMT
- Title: An Empirical Study of Pseudo-Labeling for Image-based 3D Object
Detection
- Authors: Xinzhu Ma, Yuan Meng, Yinmin Zhang, Lei Bai, Jun Hou, Shuai Yi, and
Wanli Ouyang
- Abstract summary: We investigate whether pseudo-labels can provide effective supervision for the baseline models under varying settings.
We achieve 20.23 AP for moderate level on the KITTI-3D testing set without bells and whistles, improving the baseline model by 6.03 AP.
We hope this work can provide insights for the image-based 3D detection community under a semi-supervised setting.
- Score: 72.30883544352918
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image-based 3D detection is an indispensable component of the perception
system for autonomous driving. However, it still suffers from the unsatisfying
performance, one of the main reasons for which is the limited training data.
Unfortunately, annotating the objects in the 3D space is extremely
time/resource-consuming, which makes it hard to extend the training set
arbitrarily. In this work, we focus on the semi-supervised manner and explore
the feasibility of a cheaper alternative, i.e. pseudo-labeling, to leverage the
unlabeled data. For this purpose, we conduct extensive experiments to
investigate whether the pseudo-labels can provide effective supervision for the
baseline models under varying settings. The experimental results not only
demonstrate the effectiveness of the pseudo-labeling mechanism for image-based
3D detection (e.g. under monocular setting, we achieve 20.23 AP for moderate
level on the KITTI-3D testing set without bells and whistles, improving the
baseline model by 6.03 AP), but also show several interesting and noteworthy
findings (e.g. the models trained with pseudo-labels perform better than that
trained with ground-truth annotations based on the same training data). We hope
this work can provide insights for the image-based 3D detection community under
a semi-supervised setting. The codes, pseudo-labels, and pre-trained models
will be publicly available.
Related papers
- Shelf-Supervised Cross-Modal Pre-Training for 3D Object Detection [52.66283064389691]
State-of-the-art 3D object detectors are often trained on massive labeled datasets.
Recent works demonstrate that self-supervised pre-training with unlabeled data can improve detection accuracy with limited labels.
We propose a shelf-supervised approach for generating zero-shot 3D bounding boxes from paired RGB and LiDAR data.
arXiv Detail & Related papers (2024-06-14T15:21:57Z) - View-to-Label: Multi-View Consistency for Self-Supervised 3D Object
Detection [46.077668660248534]
We propose a novel approach to self-supervise 3D object detection purely from RGB sequences alone.
Our experiments on KITTI 3D dataset demonstrate performance on par with state-of-the-art self-supervised methods.
arXiv Detail & Related papers (2023-05-29T09:30:39Z) - Augment and Criticize: Exploring Informative Samples for Semi-Supervised
Monocular 3D Object Detection [64.65563422852568]
We improve the challenging monocular 3D object detection problem with a general semi-supervised framework.
We introduce a novel, simple, yet effective Augment and Criticize' framework that explores abundant informative samples from unlabeled data.
The two new detectors, dubbed 3DSeMo_DLE and 3DSeMo_FLEX, achieve state-of-the-art results with remarkable improvements for over 3.5% AP_3D/BEV (Easy) on KITTI.
arXiv Detail & Related papers (2023-03-20T16:28:15Z) - Weakly Supervised Monocular 3D Object Detection using Multi-View
Projection and Direction Consistency [78.76508318592552]
Monocular 3D object detection has become a mainstream approach in automatic driving for its easy application.
Most current methods still rely on 3D point cloud data for labeling the ground truths used in the training phase.
We propose a new weakly supervised monocular 3D objection detection method, which can train the model with only 2D labels marked on images.
arXiv Detail & Related papers (2023-03-15T15:14:00Z) - SEAL: Self-supervised Embodied Active Learning using Exploration and 3D
Consistency [122.18108118190334]
We present a framework called Self- Embodied Embodied Active Learning (SEAL)
It utilizes perception models trained on internet images to learn an active exploration policy.
We and build utilize 3D semantic maps to learn both action and perception in a completely self-supervised manner.
arXiv Detail & Related papers (2021-12-02T06:26:38Z) - ST3D: Self-training for Unsupervised Domain Adaptation on 3D
ObjectDetection [78.71826145162092]
We present a new domain adaptive self-training pipeline, named ST3D, for unsupervised domain adaptation on 3D object detection from point clouds.
Our ST3D achieves state-of-the-art performance on all evaluated datasets and even surpasses fully supervised results on KITTI 3D object detection benchmark.
arXiv Detail & Related papers (2021-03-09T10:51:24Z) - Self-Supervised Pretraining of 3D Features on any Point-Cloud [40.26575888582241]
We present a simple self-supervised pertaining method that can work with any 3D data without 3D registration.
We evaluate our models on 9 benchmarks for object detection, semantic segmentation, and object classification, where they achieve state-of-the-art results and can outperform supervised pretraining.
arXiv Detail & Related papers (2021-01-07T18:55:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.