Are Dense Labels Always Necessary for 3D Object Detection from Point
Cloud?
- URL: http://arxiv.org/abs/2403.02818v1
- Date: Tue, 5 Mar 2024 09:38:11 GMT
- Title: Are Dense Labels Always Necessary for 3D Object Detection from Point
Cloud?
- Authors: Chenqiang Gao, Chuandong Liu, Jun Shu, Fangcen Liu, Jiang Liu, Luyu
Yang, Xinbo Gao, and Deyu Meng
- Abstract summary: Current state-of-the-art (SOTA) 3D object detection methods often require a large amount of 3D bounding box annotations for training.
We propose a novel sparsely-annotated framework, in which we just annotate one 3D object per scene.
We develop the SS3D++ method that alternatively improves 3D detector training and confident fully-annotated scene generation.
- Score: 72.40353149833109
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current state-of-the-art (SOTA) 3D object detection methods often require a
large amount of 3D bounding box annotations for training. However, collecting
such large-scale densely-supervised datasets is notoriously costly. To reduce
the cumbersome data annotation process, we propose a novel sparsely-annotated
framework, in which we just annotate one 3D object per scene. Such a sparse
annotation strategy could significantly reduce the heavy annotation burden,
while inexact and incomplete sparse supervision may severely deteriorate the
detection performance. To address this issue, we develop the SS3D++ method that
alternatively improves 3D detector training and confident fully-annotated scene
generation in a unified learning scheme. Using sparse annotations as seeds, we
progressively generate confident fully-annotated scenes based on designing a
missing-annotated instance mining module and reliable background mining module.
Our proposed method produces competitive results when compared with SOTA
weakly-supervised methods using the same or even more annotation costs.
Besides, compared with SOTA fully-supervised methods, we achieve on-par or even
better performance on the KITTI dataset with about 5x less annotation cost, and
90% of their performance on the Waymo dataset with about 15x less annotation
cost. The additional unlabeled training scenes could further boost the
performance. The code will be available at https://github.com/gaocq/SS3D2.
Related papers
- Bayesian Self-Training for Semi-Supervised 3D Segmentation [59.544558398992386]
3D segmentation is a core problem in computer vision.
densely labeling 3D point clouds to employ fully-supervised training remains too labor intensive and expensive.
Semi-supervised training provides a more practical alternative, where only a small set of labeled data is given, accompanied by a larger unlabeled set.
arXiv Detail & Related papers (2024-09-12T14:54:31Z) - ALPI: Auto-Labeller with Proxy Injection for 3D Object Detection using 2D Labels Only [5.699475977818167]
3D object detection plays a crucial role in various applications such as autonomous vehicles, robotics and augmented reality.
We propose a weakly supervised 3D annotator that relies solely on 2D bounding box annotations from images, along with size priors.
arXiv Detail & Related papers (2024-07-24T11:58:31Z) - Weakly Supervised 3D Object Detection with Multi-Stage Generalization [62.96670547848691]
We introduce BA$2$-Det, encompassing pseudo label generation and multi-stage generalization.
We develop three stages of generalization: progressing from complete to partial, static to dynamic, and close to distant.
BA$2$-Det can achieve a 20% relative improvement on the KITTI dataset.
arXiv Detail & Related papers (2023-06-08T17:58:57Z) - Weakly Supervised Monocular 3D Object Detection using Multi-View
Projection and Direction Consistency [78.76508318592552]
Monocular 3D object detection has become a mainstream approach in automatic driving for its easy application.
Most current methods still rely on 3D point cloud data for labeling the ground truths used in the training phase.
We propose a new weakly supervised monocular 3D objection detection method, which can train the model with only 2D labels marked on images.
arXiv Detail & Related papers (2023-03-15T15:14:00Z) - Semi-supervised 3D Object Detection via Adaptive Pseudo-Labeling [18.209409027211404]
3D object detection is an important task in computer vision.
Most existing methods require a large number of high-quality 3D annotations, which are expensive to collect.
We propose a novel semi-supervised framework based on pseudo-labeling for outdoor 3D object detection tasks.
arXiv Detail & Related papers (2021-08-15T02:58:43Z) - Weakly Supervised Learning of Rigid 3D Scene Flow [81.37165332656612]
We propose a data-driven scene flow estimation algorithm exploiting the observation that many 3D scenes can be explained by a collection of agents moving as rigid bodies.
We showcase the effectiveness and generalization capacity of our method on four different autonomous driving datasets.
arXiv Detail & Related papers (2021-02-17T18:58:02Z) - SESS: Self-Ensembling Semi-Supervised 3D Object Detection [138.80825169240302]
We propose SESS, a self-ensembling semi-supervised 3D object detection framework. Specifically, we design a thorough perturbation scheme to enhance generalization of the network on unlabeled and new unseen data.
Our SESS achieves competitive performance compared to the state-of-the-art fully-supervised method by using only 50% labeled data.
arXiv Detail & Related papers (2019-12-26T08:48:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.