Sample, Crop, Track: Self-Supervised Mobile 3D Object Detection for
Urban Driving LiDAR
- URL: http://arxiv.org/abs/2209.10471v1
- Date: Wed, 21 Sep 2022 16:12:46 GMT
- Title: Sample, Crop, Track: Self-Supervised Mobile 3D Object Detection for
Urban Driving LiDAR
- Authors: Sangyun Shin, Stuart Golodetz, Madhu Vankadari, Kaichen Zhou, Andrew
Markham, Niki Trigoni
- Abstract summary: We propose a new self-supervised mobile object detection approach called SCT.
This uses both motion cues and expected object sizes to improve detection performance.
We significantly outperform the state-of-the-art self-supervised mobile object detection method TCR on the KITTI tracking benchmark.
- Score: 43.971680545189756
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deep learning has led to great progress in the detection of mobile (i.e.
movement-capable) objects in urban driving scenes in recent years. Supervised
approaches typically require the annotation of large training sets; there has
thus been great interest in leveraging weakly, semi- or self-supervised methods
to avoid this, with much success. Whilst weakly and semi-supervised methods
require some annotation, self-supervised methods have used cues such as motion
to relieve the need for annotation altogether. However, a complete absence of
annotation typically degrades their performance, and ambiguities that arise
during motion grouping can inhibit their ability to find accurate object
boundaries. In this paper, we propose a new self-supervised mobile object
detection approach called SCT. This uses both motion cues and expected object
sizes to improve detection performance, and predicts a dense grid of 3D
oriented bounding boxes to improve object discovery. We significantly
outperform the state-of-the-art self-supervised mobile object detection method
TCR on the KITTI tracking benchmark, and achieve performance that is within 30%
of the fully supervised PV-RCNN++ method for IoUs <= 0.5.
Related papers
- Weakly-supervised Contrastive Learning with Quantity Prompts for Moving Infrared Small Target Detection [11.930404803127358]
Moving infrared small target detection faces huge challenges due to tiny target size and weak background contrast.<n>Currently, most existing methods are fully-supervised, heavily relying on a large number of manual target-wise annotations.<n>This paper proposes a new weakly-supervised contrastive learning (WeCoL) scheme, only requires simple target quantity prompts during model training.
arXiv Detail & Related papers (2025-07-03T09:11:31Z) - Inverse++: Vision-Centric 3D Semantic Occupancy Prediction Assisted with 3D Object Detection [11.33083039877258]
3D semantic occupancy prediction aims to forecast detailed geometric and semantic information of the surrounding environment for autonomous vehicles.
We introduce an additional 3D supervision signal by incorporating an additional 3D object detection auxiliary branch.
Our approach attains state-of-the-art results, achieving an IoU score of 31.73% and a mIoU score of 20.91%.
arXiv Detail & Related papers (2025-04-07T05:08:22Z) - Street Gaussians without 3D Object Tracker [86.62329193275916]
Existing methods rely on labor-intensive manual labeling of object poses to reconstruct dynamic objects in canonical space.
We propose a stable object tracking module by leveraging associations from 2D deep trackers within a 3D object fusion strategy.
We address inevitable tracking errors by further introducing a motion learning strategy in an implicit feature space that autonomously corrects trajectory errors and recovers missed detections.
arXiv Detail & Related papers (2024-12-07T05:49:42Z) - Cross-Cluster Shifting for Efficient and Effective 3D Object Detection
in Autonomous Driving [69.20604395205248]
We present a new 3D point-based detector model, named Shift-SSD, for precise 3D object detection in autonomous driving.
We introduce an intriguing Cross-Cluster Shifting operation to unleash the representation capacity of the point-based detector.
We conduct extensive experiments on the KITTI, runtime, and nuScenes datasets, and the results demonstrate the state-of-the-art performance of Shift-SSD.
arXiv Detail & Related papers (2024-03-10T10:36:32Z) - SeMoLi: What Moves Together Belongs Together [51.72754014130369]
We tackle semi-supervised object detection based on motion cues.
Recent results suggest that motion-based clustering methods can be used to pseudo-label instances of moving objects.
We re-think this approach and suggest that both, object detection, as well as motion-inspired pseudo-labeling, can be tackled in a data-driven manner.
arXiv Detail & Related papers (2024-02-29T18:54:53Z) - Improving Online Lane Graph Extraction by Object-Lane Clustering [106.71926896061686]
We propose an architecture and loss formulation to improve the accuracy of local lane graph estimates.
The proposed method learns to assign the objects to centerlines by considering the centerlines as cluster centers.
We show that our method can achieve significant performance improvements by using the outputs of existing 3D object detection methods.
arXiv Detail & Related papers (2023-07-20T15:21:28Z) - View-to-Label: Multi-View Consistency for Self-Supervised 3D Object
Detection [46.077668660248534]
We propose a novel approach to self-supervise 3D object detection purely from RGB sequences alone.
Our experiments on KITTI 3D dataset demonstrate performance on par with state-of-the-art self-supervised methods.
arXiv Detail & Related papers (2023-05-29T09:30:39Z) - Once Detected, Never Lost: Surpassing Human Performance in Offline LiDAR
based 3D Object Detection [50.959453059206446]
This paper aims for high-performance offline LiDAR-based 3D object detection.
We first observe that experienced human annotators annotate objects from a track-centric perspective.
We propose a high-performance offline detector in a track-centric perspective instead of the conventional object-centric perspective.
arXiv Detail & Related papers (2023-04-24T17:59:05Z) - Motion Inspired Unsupervised Perception and Prediction in Autonomous
Driving [29.731790562352344]
This paper pioneers a novel and challenging direction, i.e., training perception and prediction models to understand open-set moving objects.
Our proposed framework uses self-learned flow to trigger an automated meta labeling pipeline to achieve automatic supervision.
We show that our approach generates highly promising results in open-set 3D detection and trajectory prediction.
arXiv Detail & Related papers (2022-10-14T18:55:44Z) - Exploring Diversity-based Active Learning for 3D Object Detection in Autonomous Driving [45.405303803618]
We investigate diversity-based active learning (AL) as a potential solution to alleviate the annotation burden.
We propose a novel acquisition function that enforces spatial and temporal diversity in the selected samples.
We demonstrate the effectiveness of the proposed method on the nuScenes dataset and show that it outperforms existing AL strategies significantly.
arXiv Detail & Related papers (2022-05-16T14:21:30Z) - Time-to-Label: Temporal Consistency for Self-Supervised Monocular 3D
Object Detection [46.077668660248534]
We argue that the temporal consistency on the level of object poses, provides an important supervision signal.
Specifically, we propose a self-supervised loss which uses this consistency, in addition to render-and-compare losses.
We finetune a synthetically trained monocular 3D object detection model using the pseudo-labels that we generated on real data.
arXiv Detail & Related papers (2022-03-04T08:55:49Z) - SESS: Self-Ensembling Semi-Supervised 3D Object Detection [138.80825169240302]
We propose SESS, a self-ensembling semi-supervised 3D object detection framework. Specifically, we design a thorough perturbation scheme to enhance generalization of the network on unlabeled and new unseen data.
Our SESS achieves competitive performance compared to the state-of-the-art fully-supervised method by using only 50% labeled data.
arXiv Detail & Related papers (2019-12-26T08:48:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.