Towards Unsupervised Object Detection From LiDAR Point Clouds
- URL: http://arxiv.org/abs/2311.02007v1
- Date: Fri, 3 Nov 2023 16:12:01 GMT
- Title: Towards Unsupervised Object Detection From LiDAR Point Clouds
- Authors: Lunjun Zhang, Anqi Joyce Yang, Yuwen Xiong, Sergio Casas, Bin Yang,
Mengye Ren, Raquel Urtasun
- Abstract summary: OYSTER (Object Discovery via Spatio-Temporal Refinement) is able to detect objects in a zero-shot manner without supervised finetuning.
We propose a new planning-centric perception metric based on distance-to-collision.
- Score: 46.57452180314863
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we study the problem of unsupervised object detection from 3D
point clouds in self-driving scenes. We present a simple yet effective method
that exploits (i) point clustering in near-range areas where the point clouds
are dense, (ii) temporal consistency to filter out noisy unsupervised
detections, (iii) translation equivariance of CNNs to extend the auto-labels to
long range, and (iv) self-supervision for improving on its own. Our approach,
OYSTER (Object Discovery via Spatio-Temporal Refinement), does not impose
constraints on data collection (such as repeated traversals of the same
location), is able to detect objects in a zero-shot manner without supervised
finetuning (even in sparse, distant regions), and continues to self-improve
given more rounds of iterative self-training. To better measure model
performance in self-driving scenarios, we propose a new planning-centric
perception metric based on distance-to-collision. We demonstrate that our
unsupervised object detector significantly outperforms unsupervised baselines
on PandaSet and Argoverse 2 Sensor dataset, showing promise that
self-supervision combined with object priors can enable object discovery in the
wild. For more information, visit the project website:
https://waabi.ai/research/oyster
Related papers
- Vision-Language Guidance for LiDAR-based Unsupervised 3D Object Detection [16.09503890891102]
We propose an unsupervised 3D detection approach that operates exclusively on LiDAR point clouds.
We exploit the inherent CLI-temporal knowledge of LiDAR point clouds for clustering, tracking, as well as boxtext and label refinement.
Our approach outperforms state-of-the-art unsupervised 3D object detectors on the Open dataset.
arXiv Detail & Related papers (2024-08-07T14:14:53Z) - SeMoLi: What Moves Together Belongs Together [51.72754014130369]
We tackle semi-supervised object detection based on motion cues.
Recent results suggest that motion-based clustering methods can be used to pseudo-label instances of moving objects.
We re-think this approach and suggest that both, object detection, as well as motion-inspired pseudo-labeling, can be tackled in a data-driven manner.
arXiv Detail & Related papers (2024-02-29T18:54:53Z) - MoDAR: Using Motion Forecasting for 3D Object Detection in Point Cloud
Sequences [38.7464958249103]
We propose MoDAR, using motion forecasting outputs as a type of virtual modality, to augment LiDAR point clouds.
A fused point cloud of both raw sensor points and the virtual points can then be fed to any off-the-shelf point-cloud based 3D object detector.
arXiv Detail & Related papers (2023-06-05T19:28:19Z) - Unsupervised Adaptation from Repeated Traversals for Autonomous Driving [54.59577283226982]
Self-driving cars must generalize to the end-user's environment to operate reliably.
One potential solution is to leverage unlabeled data collected from the end-users' environments.
There is no reliable signal in the target domain to supervise the adaptation process.
We show that this simple additional assumption is sufficient to obtain a potent signal that allows us to perform iterative self-training of 3D object detectors on the target domain.
arXiv Detail & Related papers (2023-03-27T15:07:55Z) - Exploiting Playbacks in Unsupervised Domain Adaptation for 3D Object
Detection [55.12894776039135]
State-of-the-art 3D object detectors, based on deep learning, have shown promising accuracy but are prone to over-fit to domain idiosyncrasies.
We propose a novel learning approach that drastically reduces this gap by fine-tuning the detector on pseudo-labels in the target domain.
We show, on five autonomous driving datasets, that fine-tuning the detector on these pseudo-labels substantially reduces the domain gap to new driving environments.
arXiv Detail & Related papers (2021-03-26T01:18:11Z) - Detecting Invisible People [58.49425715635312]
We re-purpose tracking benchmarks and propose new metrics for the task of detecting invisible objects.
We demonstrate that current detection and tracking systems perform dramatically worse on this task.
Second, we build dynamic models that explicitly reason in 3D, making use of observations produced by state-of-the-art monocular depth estimation networks.
arXiv Detail & Related papers (2020-12-15T16:54:45Z) - Unsupervised Object Detection with LiDAR Clues [70.73881791310495]
We present the first practical method for unsupervised object detection with the aid of LiDAR clues.
In our approach, candidate object segments based on 3D point clouds are firstly generated.
Then, an iterative segment labeling process is conducted to assign segment labels and to train a segment labeling network.
The labeling process is carefully designed so as to mitigate the issue of long-tailed and open-ended distribution.
arXiv Detail & Related papers (2020-11-25T18:59:54Z) - SegVoxelNet: Exploring Semantic Context and Depth-aware Features for 3D
Vehicle Detection from Point Cloud [39.99118618229583]
We propose a unified model SegVoxelNet to address the above two problems.
A semantic context encoder is proposed to leverage the free-of-charge semantic segmentation masks in the bird's eye view.
A novel depth-aware head is designed to explicitly model the distribution differences and each part of the depth-aware head is made to focus on its own target detection range.
arXiv Detail & Related papers (2020-02-13T02:42:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.