3D3L: Deep Learned 3D Keypoint Detection and Description for LiDARs
- URL: http://arxiv.org/abs/2103.13808v1
- Date: Thu, 25 Mar 2021 13:08:07 GMT
- Title: 3D3L: Deep Learned 3D Keypoint Detection and Description for LiDARs
- Authors: Dominc Streiff, Lukas Bernreiter, Florian Tschopp, Marius Fehr, Roland
Siegwart
- Abstract summary: In this publication, we use a state-of-the-art 2D feature network as a basis for 3D3L, exploiting both intensity and depth of LiDAR range images.
Our results show that these keypoints and descriptors extracted from LiDAR scan images outperform state-of-the-art on different benchmark metrics.
- Score: 25.73598441491818
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the advent of powerful, light-weight 3D LiDARs, they have become the
hearth of many navigation and SLAM algorithms on various autonomous systems.
Pointcloud registration methods working with unstructured pointclouds such as
ICP are often computationally expensive or require a good initial guess.
Furthermore, 3D feature-based registration methods have never quite reached the
robustness of 2D methods in visual SLAM. With the continuously increasing
resolution of LiDAR range images, these 2D methods not only become applicable
but should exploit the illumination-independent modalities that come with it,
such as depth and intensity. In visual SLAM, deep learned 2D features and
descriptors perform exceptionally well compared to traditional methods. In this
publication, we use a state-of-the-art 2D feature network as a basis for 3D3L,
exploiting both intensity and depth of LiDAR range images to extract powerful
3D features. Our results show that these keypoints and descriptors extracted
from LiDAR scan images outperform state-of-the-art on different benchmark
metrics and allow for robust scan-to-scan alignment as well as global
localization.
Related papers
- Weakly Supervised Monocular 3D Detection with a Single-View Image [58.57978772009438]
Monocular 3D detection aims for precise 3D object localization from a single-view image.
We propose SKD-WM3D, a weakly supervised monocular 3D detection framework.
We show that SKD-WM3D surpasses the state-of-the-art clearly and is even on par with many fully supervised methods.
arXiv Detail & Related papers (2024-02-29T13:26:47Z) - OpenSight: A Simple Open-Vocabulary Framework for LiDAR-Based Object
Detection [41.24059083441953]
OpenSight is a more advanced 2D-3D modeling framework for LiDAR-based open-vocabulary detection.
Our method establishes state-of-the-art open-vocabulary performance on widely used 3D detection benchmarks.
arXiv Detail & Related papers (2023-12-12T07:49:30Z) - MinkLoc3D-SI: 3D LiDAR place recognition with sparse convolutions,
spherical coordinates, and intensity [1.1549572298362785]
We introduce MinkLoc3D-SI, a sparse convolution-based solution that processes the intensity of 3D LiDAR measurements.
Our experiments show improved results on single scans from 3D LiDARs and great generalization ability.
MinkLoc3D-SI is suited for single scans obtained from a 3D LiDAR, making it applicable in autonomous vehicles.
arXiv Detail & Related papers (2021-12-13T10:21:34Z) - LIGA-Stereo: Learning LiDAR Geometry Aware Representations for
Stereo-based 3D Detector [80.7563981951707]
We propose LIGA-Stereo to learn stereo-based 3D detectors under the guidance of high-level geometry-aware representations of LiDAR-based detection models.
Compared with the state-of-the-art stereo detector, our method has improved the 3D detection performance of cars, pedestrians, cyclists by 10.44%, 5.69%, 5.97% mAP respectively.
arXiv Detail & Related papers (2021-08-18T17:24:40Z) - 3D-to-2D Distillation for Indoor Scene Parsing [78.36781565047656]
We present a new approach that enables us to leverage 3D features extracted from large-scale 3D data repository to enhance 2D features extracted from RGB images.
First, we distill 3D knowledge from a pretrained 3D network to supervise a 2D network to learn simulated 3D features from 2D features during the training.
Second, we design a two-stage dimension normalization scheme to calibrate the 2D and 3D features for better integration.
Third, we design a semantic-aware adversarial training model to extend our framework for training with unpaired 3D data.
arXiv Detail & Related papers (2021-04-06T02:22:24Z) - YOLOStereo3D: A Step Back to 2D for Efficient Stereo 3D Detection [6.5702792909006735]
YOLOStereo3D is trained on one single GPU and runs at more than ten fps.
It demonstrates performance comparable to state-of-the-art stereo 3D detection frameworks without usage of LiDAR data.
arXiv Detail & Related papers (2021-03-17T03:43:54Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - RoIFusion: 3D Object Detection from LiDAR and Vision [7.878027048763662]
We propose a novel fusion algorithm by projecting a set of 3D Region of Interests (RoIs) from the point clouds to the 2D RoIs of the corresponding the images.
Our approach achieves state-of-the-art performance on the KITTI 3D object detection challenging benchmark.
arXiv Detail & Related papers (2020-09-09T20:23:27Z) - End-to-End Pseudo-LiDAR for Image-Based 3D Object Detection [62.34374949726333]
Pseudo-LiDAR (PL) has led to a drastic reduction in the accuracy gap between methods based on LiDAR sensors and those based on cheap stereo cameras.
PL combines state-of-the-art deep neural networks for 3D depth estimation with those for 3D object detection by converting 2D depth map outputs to 3D point cloud inputs.
We introduce a new framework based on differentiable Change of Representation (CoR) modules that allow the entire PL pipeline to be trained end-to-end.
arXiv Detail & Related papers (2020-04-07T02:18:38Z) - DSGN: Deep Stereo Geometry Network for 3D Object Detection [79.16397166985706]
There is a large performance gap between image-based and LiDAR-based 3D object detectors.
Our method, called Deep Stereo Geometry Network (DSGN), significantly reduces this gap.
For the first time, we provide a simple and effective one-stage stereo-based 3D detection pipeline.
arXiv Detail & Related papers (2020-01-10T11:44:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.