Self-Supervised 3D Monocular Object Detection by Recycling Bounding
Boxes
- URL: http://arxiv.org/abs/2206.12738v1
- Date: Sat, 25 Jun 2022 21:48:43 GMT
- Title: Self-Supervised 3D Monocular Object Detection by Recycling Bounding
Boxes
- Authors: Sugirtha T, Sridevi M, Khailash Santhakumar, Hao Liu, B Ravi Kiran,
Thomas Gauthier, Senthil Yogamani
- Abstract summary: The paper studies the application of established self-supervised bounding box recycling by labeling random windows as the pretext task.
We demonstrate improvements of between 2-3 % in mAP 3D and 0.9-1.5 % BEV scores using SSL over the baseline scores.
- Score: 3.3299316770988625
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modern object detection architectures are moving towards employing
self-supervised learning (SSL) to improve performance detection with related
pretext tasks. Pretext tasks for monocular 3D object detection have not yet
been explored yet in literature. The paper studies the application of
established self-supervised bounding box recycling by labeling random windows
as the pretext task. The classifier head of the 3D detector is trained to
classify random windows containing different proportions of the ground truth
objects, thus handling the foreground-background imbalance. We evaluate the
pretext task using the RTM3D detection model as baseline, with and without the
application of data augmentation. We demonstrate improvements of between 2-3 %
in mAP 3D and 0.9-1.5 % BEV scores using SSL over the baseline scores. We
propose the inverse class frequency re-weighted (ICFW) mAP score that
highlights improvements in detection for low frequency classes in a class
imbalanced dataset with long tails. We demonstrate improvements in ICFW both
mAP 3D and BEV scores to take into account the class imbalance in the KITTI
validation dataset. We see 4-5 % increase in ICFW metric with the pretext task.
Related papers
- Uncertainty Estimation for 3D Object Detection via Evidential Learning [63.61283174146648]
We introduce a framework for quantifying uncertainty in 3D object detection by leveraging an evidential learning loss on Bird's Eye View representations in the 3D detector.
We demonstrate both the efficacy and importance of these uncertainty estimates on identifying out-of-distribution scenes, poorly localized objects, and missing (false negative) detections.
arXiv Detail & Related papers (2024-10-31T13:13:32Z) - FILP-3D: Enhancing 3D Few-shot Class-incremental Learning with
Pre-trained Vision-Language Models [62.663113296987085]
Few-shot class-incremental learning aims to mitigate the catastrophic forgetting issue when a model is incrementally trained on limited data.
We introduce two novel components: the Redundant Feature Eliminator (RFE) and the Spatial Noise Compensator (SNC)
Considering the imbalance in existing 3D datasets, we also propose new evaluation metrics that offer a more nuanced assessment of a 3D FSCIL model.
arXiv Detail & Related papers (2023-12-28T14:52:07Z) - Augment and Criticize: Exploring Informative Samples for Semi-Supervised
Monocular 3D Object Detection [64.65563422852568]
We improve the challenging monocular 3D object detection problem with a general semi-supervised framework.
We introduce a novel, simple, yet effective Augment and Criticize' framework that explores abundant informative samples from unlabeled data.
The two new detectors, dubbed 3DSeMo_DLE and 3DSeMo_FLEX, achieve state-of-the-art results with remarkable improvements for over 3.5% AP_3D/BEV (Easy) on KITTI.
arXiv Detail & Related papers (2023-03-20T16:28:15Z) - Delving into the Pre-training Paradigm of Monocular 3D Object Detection [10.07932482761621]
We study the pre-training paradigm for monocular 3D object detection (M3OD)
We propose several strategies to further improve this baseline, which mainly include target guided semi-dense depth estimation, keypoint-aware 2D object detection, and class-level loss adjustment.
Combining all the developed techniques, the obtained pre-training framework produces pre-trained backbones that improve M3OD performance significantly on the KITTI-3D and nuScenes benchmarks.
arXiv Detail & Related papers (2022-06-08T03:01:13Z) - ST3D++: Denoised Self-training for Unsupervised Domain Adaptation on 3D
Object Detection [78.71826145162092]
We present a self-training method, named ST3D++, with a holistic pseudo label denoising pipeline for unsupervised domain adaptation on 3D object detection.
We equip the pseudo label generation process with a hybrid quality-aware triplet memory to improve the quality and stability of generated pseudo labels.
In the model training stage, we propose a source data assisted training strategy and a curriculum data augmentation policy.
arXiv Detail & Related papers (2021-08-15T07:49:06Z) - Exploring 2D Data Augmentation for 3D Monocular Object Detection [0.2936007114555107]
Many standard 2D object detection data augmentation techniques do not extend to 3D box.
We propose two novel augmentations for monocular 3D detection without a requirement for novel view synthesis.
arXiv Detail & Related papers (2021-04-21T22:43:42Z) - ST3D: Self-training for Unsupervised Domain Adaptation on 3D
ObjectDetection [78.71826145162092]
We present a new domain adaptive self-training pipeline, named ST3D, for unsupervised domain adaptation on 3D object detection from point clouds.
Our ST3D achieves state-of-the-art performance on all evaluated datasets and even surpasses fully supervised results on KITTI 3D object detection benchmark.
arXiv Detail & Related papers (2021-03-09T10:51:24Z) - 3DIoUMatch: Leveraging IoU Prediction for Semi-Supervised 3D Object
Detection [76.42897462051067]
3DIoUMatch is a novel semi-supervised method for 3D object detection applicable to both indoor and outdoor scenes.
We leverage a teacher-student mutual learning framework to propagate information from the labeled to the unlabeled train set in the form of pseudo-labels.
Our method consistently improves state-of-the-art methods on both ScanNet and SUN-RGBD benchmarks by significant margins under all label ratios.
arXiv Detail & Related papers (2020-12-08T11:06:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.