HPointLoc: Point-based Indoor Place Recognition using Synthetic RGB-D
Images
- URL: http://arxiv.org/abs/2212.14649v1
- Date: Fri, 30 Dec 2022 12:20:56 GMT
- Title: HPointLoc: Point-based Indoor Place Recognition using Synthetic RGB-D
Images
- Authors: Dmitry Yudin, Yaroslav Solomentsev, Ruslan Musaev, Aleksei Staroverov,
Aleksandr I. Panov
- Abstract summary: We present a novel dataset named as HPointLoc, specially designed for exploring capabilities of visual place recognition in indoor environment.
The dataset is based on the popular Habitat simulator, in which it is possible to generate indoor scenes using both own sensor data and open datasets.
- Score: 58.720142291102135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a novel dataset named as HPointLoc, specially designed for
exploring capabilities of visual place recognition in indoor environment and
loop detection in simultaneous localization and mapping. The loop detection
sub-task is especially relevant when a robot with an on-board RGB-D camera can
drive past the same place (``Point") at different angles. The dataset is based
on the popular Habitat simulator, in which it is possible to generate
photorealistic indoor scenes using both own sensor data and open datasets, such
as Matterport3D. To study the main stages of solving the place recognition
problem on the HPointLoc dataset, we proposed a new modular approach named as
PNTR. It first performs an image retrieval with the Patch-NetVLAD method, then
extracts keypoints and matches them using R2D2, LoFTR or SuperPoint with
SuperGlue, and finally performs a camera pose optimization step with TEASER++.
Such a solution to the place recognition problem has not been previously
studied in existing publications. The PNTR approach has shown the best quality
metrics on the HPointLoc dataset and has a high potential for real use in
localization systems for unmanned vehicles. The proposed dataset and framework
are publicly available: https://github.com/metra4ok/HPointLoc.
Related papers
- PointRegGPT: Boosting 3D Point Cloud Registration using Generative Point-Cloud Pairs for Training [90.06520673092702]
We present PointRegGPT, boosting 3D point cloud registration using generative point-cloud pairs for training.
To our knowledge, this is the first generative approach that explores realistic data generation for indoor point cloud registration.
arXiv Detail & Related papers (2024-07-19T06:29:57Z) - V-DETR: DETR with Vertex Relative Position Encoding for 3D Object
Detection [73.37781484123536]
We introduce a highly performant 3D object detector for point clouds using the DETR framework.
To address the limitation, we introduce a novel 3D Relative Position (3DV-RPE) method.
We show exceptional results on the challenging ScanNetV2 benchmark.
arXiv Detail & Related papers (2023-08-08T17:14:14Z) - Neural Implicit Dense Semantic SLAM [83.04331351572277]
We propose a novel RGBD vSLAM algorithm that learns a memory-efficient, dense 3D geometry, and semantic segmentation of an indoor scene in an online manner.
Our pipeline combines classical 3D vision-based tracking and loop closing with neural fields-based mapping.
Our proposed algorithm can greatly enhance scene perception and assist with a range of robot control problems.
arXiv Detail & Related papers (2023-04-27T23:03:52Z) - From Voxel to Point: IoU-guided 3D Object Detection for Point Cloud with
Voxel-to-Point Decoder [79.39041453836793]
We present an Intersection-over-Union (IoU) guided two-stage 3D object detector with a voxel-to-point decoder.
We propose a residual voxel-to-point decoder to extract the point features in addition to the map-view features from the voxel based Region Proposal Network (RPN)
We propose a simple and efficient method to align the estimated IoUs to the refined proposal boxes as a more relevant localization confidence.
arXiv Detail & Related papers (2021-08-08T14:30:13Z) - Large-scale Localization Datasets in Crowded Indoor Spaces [23.071409425965772]
We introduce 5 new indoor datasets for visual localization in challenging real-world environments.
They were captured in a large shopping mall and a large metro station in Seoul, South Korea.
In order to obtain accurate ground truth camera poses, we developed a robust LiDAR SLAM.
arXiv Detail & Related papers (2021-05-19T06:20:49Z) - Soft Expectation and Deep Maximization for Image Feature Detection [68.8204255655161]
We propose SEDM, an iterative semi-supervised learning process that flips the question and first looks for repeatable 3D points, then trains a detector to localize them in image space.
Our results show that this new model trained using SEDM is able to better localize the underlying 3D points in a scene.
arXiv Detail & Related papers (2021-04-21T00:35:32Z) - Robust Place Recognition using an Imaging Lidar [45.37172889338924]
We propose a methodology for robust, real-time place recognition using an imaging lidar.
Our method is truly-invariant and can tackle reverse revisiting and upside-down revisiting.
arXiv Detail & Related papers (2021-03-03T01:08:31Z) - LCD -- Line Clustering and Description for Place Recognition [29.053923938306323]
We introduce a novel learning-based approach to place recognition, using RGB-D cameras and line clusters as visual and geometric features.
We present a neural network architecture based on the attention mechanism for frame-wise line clustering.
A similar neural network is used for the description of these clusters with a compact embedding of 128 floating point numbers.
arXiv Detail & Related papers (2020-10-21T09:52:47Z) - Distinctive 3D local deep descriptors [2.512827436728378]
Point cloud patches are extracted, canonicalised with respect to their estimated local reference frame and encoded by a PointNet-based deep neural network.
We evaluate and compare DIPs against alternative hand-crafted and deep descriptors on several datasets consisting of point clouds reconstructed using different sensors.
arXiv Detail & Related papers (2020-09-01T06:25:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.