LinK3D: Linear Keypoints Representation for 3D LiDAR Point Cloud
- URL: http://arxiv.org/abs/2206.05927v3
- Date: Wed, 10 Jan 2024 15:36:16 GMT
- Title: LinK3D: Linear Keypoints Representation for 3D LiDAR Point Cloud
- Authors: Yunge Cui, Yinlong Zhang, Jiahua Dong, Haibo Sun, Xieyuanli Chen and
Feng Zhu
- Abstract summary: We propose a novel 3D feature representation method: Linear Keypoints representation for 3D LiDAR point cloud, called LinK3D.
LinK3D shows excellent real-time performance, faster than the sensor frame rate at 10 Hz of a typical rotating LiDAR sensor.
Our method can be extended to LiDAR odometry task, and shows good scalability.
- Score: 18.942933892804028
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Feature extraction and matching are the basic parts of many robotic vision
tasks, such as 2D or 3D object detection, recognition, and registration. As is
known, 2D feature extraction and matching have already achieved great success.
Unfortunately, in the field of 3D, the current methods may fail to support the
extensive application of 3D LiDAR sensors in robotic vision tasks due to their
poor descriptiveness and inefficiency. To address this limitation, we propose a
novel 3D feature representation method: Linear Keypoints representation for 3D
LiDAR point cloud, called LinK3D. The novelty of LinK3D lies in that it fully
considers the characteristics (such as the sparsity and complexity) of LiDAR
point clouds and represents the keypoint with its robust neighbor keypoints,
which provide strong constraints in the description of the keypoint. The
proposed LinK3D has been evaluated on three public datasets, and the
experimental results show that our method achieves great matching performance.
More importantly, LinK3D also shows excellent real-time performance, faster
than the sensor frame rate at 10 Hz of a typical rotating LiDAR sensor. LinK3D
only takes an average of 30 milliseconds to extract features from the point
cloud collected by a 64-beam LiDAR and takes merely about 20 milliseconds to
match two LiDAR scans when executed on a computer with an Intel Core i7
processor. Moreover, our method can be extended to LiDAR odometry task, and
shows good scalability. We release the implementation of our method at
https://github.com/YungeCui/LinK3D.
Related papers
- Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - Weakly Supervised 3D Object Detection via Multi-Level Visual Guidance [72.6809373191638]
We propose a framework to study how to leverage constraints between 2D and 3D domains without requiring any 3D labels.
Specifically, we design a feature-level constraint to align LiDAR and image features based on object-aware regions.
Second, the output-level constraint is developed to enforce the overlap between 2D and projected 3D box estimations.
Third, the training-level constraint is utilized by producing accurate and consistent 3D pseudo-labels that align with the visual data.
arXiv Detail & Related papers (2023-12-12T18:57:25Z) - Sparse2Dense: Learning to Densify 3D Features for 3D Object Detection [85.08249413137558]
LiDAR-produced point clouds are the major source for most state-of-the-art 3D object detectors.
Small, distant, and incomplete objects with sparse or few points are often hard to detect.
We present Sparse2Dense, a new framework to efficiently boost 3D detection performance by learning to densify point clouds in latent space.
arXiv Detail & Related papers (2022-11-23T16:01:06Z) - MinkLoc3D-SI: 3D LiDAR place recognition with sparse convolutions,
spherical coordinates, and intensity [1.1549572298362785]
We introduce MinkLoc3D-SI, a sparse convolution-based solution that processes the intensity of 3D LiDAR measurements.
Our experiments show improved results on single scans from 3D LiDARs and great generalization ability.
MinkLoc3D-SI is suited for single scans obtained from a 3D LiDAR, making it applicable in autonomous vehicles.
arXiv Detail & Related papers (2021-12-13T10:21:34Z) - 3D3L: Deep Learned 3D Keypoint Detection and Description for LiDARs [25.73598441491818]
In this publication, we use a state-of-the-art 2D feature network as a basis for 3D3L, exploiting both intensity and depth of LiDAR range images.
Our results show that these keypoints and descriptors extracted from LiDAR scan images outperform state-of-the-art on different benchmark metrics.
arXiv Detail & Related papers (2021-03-25T13:08:07Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - RoIFusion: 3D Object Detection from LiDAR and Vision [7.878027048763662]
We propose a novel fusion algorithm by projecting a set of 3D Region of Interests (RoIs) from the point clouds to the 2D RoIs of the corresponding the images.
Our approach achieves state-of-the-art performance on the KITTI 3D object detection challenging benchmark.
arXiv Detail & Related papers (2020-09-09T20:23:27Z) - End-to-End Pseudo-LiDAR for Image-Based 3D Object Detection [62.34374949726333]
Pseudo-LiDAR (PL) has led to a drastic reduction in the accuracy gap between methods based on LiDAR sensors and those based on cheap stereo cameras.
PL combines state-of-the-art deep neural networks for 3D depth estimation with those for 3D object detection by converting 2D depth map outputs to 3D point cloud inputs.
We introduce a new framework based on differentiable Change of Representation (CoR) modules that allow the entire PL pipeline to be trained end-to-end.
arXiv Detail & Related papers (2020-04-07T02:18:38Z) - D3Feat: Joint Learning of Dense Detection and Description of 3D Local
Features [51.04841465193678]
We leverage a 3D fully convolutional network for 3D point clouds.
We propose a novel and practical learning mechanism that densely predicts both a detection score and a description feature for each 3D point.
Our method achieves state-of-the-art results in both indoor and outdoor scenarios.
arXiv Detail & Related papers (2020-03-06T12:51:09Z) - RTM3D: Real-time Monocular 3D Detection from Object Keypoints for
Autonomous Driving [26.216609821525676]
Most successful 3D detectors take the projection constraint from the 3D bounding box to the 2D box as an important component.
Our method predicts the nine perspective keypoints of a 3D bounding box in image space, and then utilize the geometric relationship of 3D and 2D perspectives to recover the dimension, location, and orientation in 3D space.
Our method is the first real-time system for monocular image 3D detection while achieves state-of-the-art performance on the KITTI benchmark.
arXiv Detail & Related papers (2020-01-10T08:29:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.