Monocular 3D Object Detection with LiDAR Guided Semi Supervised Active
Learning
- URL: http://arxiv.org/abs/2307.08415v1
- Date: Mon, 17 Jul 2023 11:55:27 GMT
- Title: Monocular 3D Object Detection with LiDAR Guided Semi Supervised Active
Learning
- Authors: Aral Hekimoglu, Michael Schmidt, Alvaro Marcos-Ramiro
- Abstract summary: We propose a novel semi-supervised active learning (SSAL) framework for monocular 3D object detection with LiDAR guidance (MonoLiG)
We utilize LiDAR to guide the data selection and training of monocular 3D detectors without introducing any overhead in the inference phase.
Our training strategy attains the top place in KITTI 3D and birds-eye-view (BEV) monocular object detection official benchmarks by improving the BEV Average Precision (AP) by 2.02.
- Score: 2.16117348324501
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We propose a novel semi-supervised active learning (SSAL) framework for
monocular 3D object detection with LiDAR guidance (MonoLiG), which leverages
all modalities of collected data during model development. We utilize LiDAR to
guide the data selection and training of monocular 3D detectors without
introducing any overhead in the inference phase. During training, we leverage
the LiDAR teacher, monocular student cross-modal framework from semi-supervised
learning to distill information from unlabeled data as pseudo-labels. To handle
the differences in sensor characteristics, we propose a data noise-based
weighting mechanism to reduce the effect of propagating noise from LiDAR
modality to monocular. For selecting which samples to label to improve the
model performance, we propose a sensor consistency-based selection score that
is also coherent with the training objective. Extensive experimental results on
KITTI and Waymo datasets verify the effectiveness of our proposed framework. In
particular, our selection strategy consistently outperforms state-of-the-art
active learning baselines, yielding up to 17% better saving rate in labeling
costs. Our training strategy attains the top place in KITTI 3D and
birds-eye-view (BEV) monocular object detection official benchmarks by
improving the BEV Average Precision (AP) by 2.02.
Related papers
- Approaching Outside: Scaling Unsupervised 3D Object Detection from 2D Scene [22.297964850282177]
We propose LiDAR-2D Self-paced Learning (LiSe) for unsupervised 3D detection.
RGB images serve as a valuable complement to LiDAR data, offering precise 2D localization cues.
Our framework devises a self-paced learning pipeline that incorporates adaptive sampling and weak model aggregation strategies.
arXiv Detail & Related papers (2024-07-11T14:58:49Z) - Multi-Modal Data-Efficient 3D Scene Understanding for Autonomous Driving [58.16024314532443]
We introduce LaserMix++, a framework that integrates laser beam manipulations from disparate LiDAR scans and incorporates LiDAR-camera correspondences to assist data-efficient learning.
Results demonstrate that LaserMix++ outperforms fully supervised alternatives, achieving comparable accuracy with five times fewer annotations.
This substantial advancement underscores the potential of semi-supervised approaches in reducing the reliance on extensive labeled data in LiDAR-based 3D scene understanding systems.
arXiv Detail & Related papers (2024-05-08T17:59:53Z) - ODM3D: Alleviating Foreground Sparsity for Semi-Supervised Monocular 3D
Object Detection [15.204935788297226]
ODM3D framework entails cross-modal knowledge distillation at various levels to inject LiDAR-domain knowledge into a monocular detector during training.
By identifying foreground sparsity as the main culprit behind existing methods' suboptimal training, we exploit the precise localisation information embedded in LiDAR points.
Our method ranks 1st in both KITTI validation and test benchmarks, significantly surpassing all existing monocular methods, supervised or semi-supervised.
arXiv Detail & Related papers (2023-10-28T07:12:09Z) - Augment and Criticize: Exploring Informative Samples for Semi-Supervised
Monocular 3D Object Detection [64.65563422852568]
We improve the challenging monocular 3D object detection problem with a general semi-supervised framework.
We introduce a novel, simple, yet effective Augment and Criticize' framework that explores abundant informative samples from unlabeled data.
The two new detectors, dubbed 3DSeMo_DLE and 3DSeMo_FLEX, achieve state-of-the-art results with remarkable improvements for over 3.5% AP_3D/BEV (Easy) on KITTI.
arXiv Detail & Related papers (2023-03-20T16:28:15Z) - LiDAL: Inter-frame Uncertainty Based Active Learning for 3D LiDAR
Semantic Segmentation [36.69417080183985]
LiDAL is a novel active learning method for 3D LiDAR semantic segmentation.
We propose LiDAL by exploiting inter-frame uncertainty among LiDAR frames.
We achieve 95% of the performance of fully supervised learning with less than 5% of annotations.
arXiv Detail & Related papers (2022-11-11T04:47:33Z) - Semi-supervised 3D Object Detection with Proficient Teachers [114.54835359657707]
Dominated point cloud-based 3D object detectors in autonomous driving scenarios rely heavily on the huge amount of accurately labeled samples.
Pseudo-Labeling methodology is commonly used for SSL frameworks, however, the low-quality predictions from the teacher model have seriously limited its performance.
We propose a new Pseudo-Labeling framework for semi-supervised 3D object detection, by enhancing the teacher model to a proficient one with several necessary designs.
arXiv Detail & Related papers (2022-07-26T04:54:03Z) - Boosting 3D Object Detection by Simulating Multimodality on Point Clouds [51.87740119160152]
This paper presents a new approach to boost a single-modality (LiDAR) 3D object detector by teaching it to simulate features and responses that follow a multi-modality (LiDAR-image) detector.
The approach needs LiDAR-image data only when training the single-modality detector, and once well-trained, it only needs LiDAR data at inference.
Experimental results on the nuScenes dataset show that our approach outperforms all SOTA LiDAR-only 3D detectors.
arXiv Detail & Related papers (2022-06-30T01:44:30Z) - Open-Set Semi-Supervised Learning for 3D Point Cloud Understanding [62.17020485045456]
It is commonly assumed in semi-supervised learning (SSL) that the unlabeled data are drawn from the same distribution as that of the labeled ones.
We propose to selectively utilize unlabeled data through sample weighting, so that only conducive unlabeled data would be prioritized.
arXiv Detail & Related papers (2022-05-02T16:09:17Z) - MonoDistill: Learning Spatial Features for Monocular 3D Object Detection [80.74622486604886]
We propose a simple and effective scheme to introduce the spatial information from LiDAR signals to the monocular 3D detectors.
We use the resulting data to train a 3D detector with the same architecture as the baseline model.
Experimental results show that the proposed method can significantly boost the performance of the baseline model.
arXiv Detail & Related papers (2022-01-26T09:21:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.