MinkLoc3D-SI: 3D LiDAR place recognition with sparse convolutions,
spherical coordinates, and intensity
- URL: http://arxiv.org/abs/2112.06539v1
- Date: Mon, 13 Dec 2021 10:21:34 GMT
- Title: MinkLoc3D-SI: 3D LiDAR place recognition with sparse convolutions,
spherical coordinates, and intensity
- Authors: Kamil \.Zywanowski, Adam Banaszczyk, Micha{\l} R. Nowicki, and Jacek
Komorowski
- Abstract summary: We introduce MinkLoc3D-SI, a sparse convolution-based solution that processes the intensity of 3D LiDAR measurements.
Our experiments show improved results on single scans from 3D LiDARs and great generalization ability.
MinkLoc3D-SI is suited for single scans obtained from a 3D LiDAR, making it applicable in autonomous vehicles.
- Score: 1.1549572298362785
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: The 3D LiDAR place recognition aims to estimate a coarse localization in a
previously seen environment based on a single scan from a rotating 3D LiDAR
sensor. The existing solutions to this problem include hand-crafted point cloud
descriptors (e.g., ScanContext, M2DP, LiDAR IRIS) and deep learning-based
solutions (e.g., PointNetVLAD, PCAN, LPDNet, DAGC, MinkLoc3D), which are often
only evaluated on accumulated 2D scans from the Oxford RobotCar dataset. We
introduce MinkLoc3D-SI, a sparse convolution-based solution that utilizes
spherical coordinates of 3D points and processes the intensity of 3D LiDAR
measurements, improving the performance when a single 3D LiDAR scan is used.
Our method integrates the improvements typical for hand-crafted descriptors
(like ScanContext) with the most efficient 3D sparse convolutions (MinkLoc3D).
Our experiments show improved results on single scans from 3D LiDARs (USyd
Campus dataset) and great generalization ability (KITTI dataset). Using
intensity information on accumulated 2D scans (RobotCar Intensity dataset)
improves the performance, even though spherical representation doesn't produce
a noticeable improvement. As a result, MinkLoc3D-SI is suited for single scans
obtained from a 3D LiDAR, making it applicable in autonomous vehicles.
Related papers
- DeepMIF: Deep Monotonic Implicit Fields for Large-Scale LiDAR 3D Mapping [46.80755234561584]
Recent learning-based methods integrate neural implicit representations and optimizable feature grids to approximate surfaces of 3D scenes.
In this work we depart from fitting LiDAR data exactly, instead letting the network optimize a non-metric monotonic implicit field defined in 3D space.
Our algorithm achieves high-quality dense 3D mapping performance as captured by multiple quantitative and perceptual measures and visual results obtained for Mai City, Newer College, and KITTI benchmarks.
arXiv Detail & Related papers (2024-03-26T09:58:06Z) - Sparse2Dense: Learning to Densify 3D Features for 3D Object Detection [85.08249413137558]
LiDAR-produced point clouds are the major source for most state-of-the-art 3D object detectors.
Small, distant, and incomplete objects with sparse or few points are often hard to detect.
We present Sparse2Dense, a new framework to efficiently boost 3D detection performance by learning to densify point clouds in latent space.
arXiv Detail & Related papers (2022-11-23T16:01:06Z) - Frustum Fusion: Pseudo-LiDAR and LiDAR Fusion for 3D Detection [0.0]
We propose a novel data fusion algorithm to combine accurate point clouds with dense but less accurate point clouds obtained from stereo pairs.
We train multiple 3D object detection methods and show that our fusion strategy consistently improves the performance of detectors.
arXiv Detail & Related papers (2021-11-08T19:29:59Z) - FGR: Frustum-Aware Geometric Reasoning for Weakly Supervised 3D Vehicle
Detection [81.79171905308827]
We propose frustum-aware geometric reasoning (FGR) to detect vehicles in point clouds without any 3D annotations.
Our method consists of two stages: coarse 3D segmentation and 3D bounding box estimation.
It is able to accurately detect objects in 3D space with only 2D bounding boxes and sparse point clouds.
arXiv Detail & Related papers (2021-05-17T07:29:55Z) - 3D-to-2D Distillation for Indoor Scene Parsing [78.36781565047656]
We present a new approach that enables us to leverage 3D features extracted from large-scale 3D data repository to enhance 2D features extracted from RGB images.
First, we distill 3D knowledge from a pretrained 3D network to supervise a 2D network to learn simulated 3D features from 2D features during the training.
Second, we design a two-stage dimension normalization scheme to calibrate the 2D and 3D features for better integration.
Third, we design a semantic-aware adversarial training model to extend our framework for training with unpaired 3D data.
arXiv Detail & Related papers (2021-04-06T02:22:24Z) - 3D3L: Deep Learned 3D Keypoint Detection and Description for LiDARs [25.73598441491818]
In this publication, we use a state-of-the-art 2D feature network as a basis for 3D3L, exploiting both intensity and depth of LiDAR range images.
Our results show that these keypoints and descriptors extracted from LiDAR scan images outperform state-of-the-art on different benchmark metrics.
arXiv Detail & Related papers (2021-03-25T13:08:07Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - D3Feat: Joint Learning of Dense Detection and Description of 3D Local
Features [51.04841465193678]
We leverage a 3D fully convolutional network for 3D point clouds.
We propose a novel and practical learning mechanism that densely predicts both a detection score and a description feature for each 3D point.
Our method achieves state-of-the-art results in both indoor and outdoor scenarios.
arXiv Detail & Related papers (2020-03-06T12:51:09Z) - DSGN: Deep Stereo Geometry Network for 3D Object Detection [79.16397166985706]
There is a large performance gap between image-based and LiDAR-based 3D object detectors.
Our method, called Deep Stereo Geometry Network (DSGN), significantly reduces this gap.
For the first time, we provide a simple and effective one-stage stereo-based 3D detection pipeline.
arXiv Detail & Related papers (2020-01-10T11:44:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.