LDL: Line Distance Functions for Panoramic Localization
- URL: http://arxiv.org/abs/2308.13989v1
- Date: Sun, 27 Aug 2023 02:57:07 GMT
- Title: LDL: Line Distance Functions for Panoramic Localization
- Authors: Junho Kim, Changwoon Choi, Hojun Jang, Young Min Kim
- Abstract summary: We introduce LDL, an algorithm that localizes a panorama to a 3D map using line segments.
Our method effectively observes the holistic distribution of lines within panoramic images and 3D maps.
- Score: 22.46846444866008
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce LDL, a fast and robust algorithm that localizes a panorama to a
3D map using line segments. LDL focuses on the sparse structural information of
lines in the scene, which is robust to illumination changes and can potentially
enable efficient computation. While previous line-based localization approaches
tend to sacrifice accuracy or computation time, our method effectively observes
the holistic distribution of lines within panoramic images and 3D maps.
Specifically, LDL matches the distribution of lines with 2D and 3D line
distance functions, which are further decomposed along principal directions of
lines to increase the expressiveness. The distance functions provide coarse
pose estimates by comparing the distributional information, where the poses are
further optimized using conventional local feature matching. As our pipeline
solely leverages line geometry and local features, it does not require costly
additional training of line-specific features or correspondence matching.
Nevertheless, our method demonstrates robust performance on challenging
scenarios including object layout changes, illumination shifts, and large-scale
scenes, while exhibiting fast pose search terminating within a matter of
milliseconds. We thus expect our method to serve as a practical solution for
line-based localization, and complement the well-established point-based
paradigm. The code for LDL is available through the following link:
https://github.com/82magnolia/panoramic-localization.
Related papers
- Fully Geometric Panoramic Localization [16.200889977514862]
We introduce a lightweight and accurate localization method that only utilizes the geometry of 2D-3D lines.
Given a pre-captured 3D map, our approach localizes a panorama image, taking advantage of the holistic 360 view.
Our fully geometric approach does not involve extensive parameter tuning or neural network training, making it a practical algorithm that can be readily deployed in the real world.
arXiv Detail & Related papers (2024-03-29T01:07:20Z) - Representing 3D sparse map points and lines for camera relocalization [1.2974519529978974]
We show how a lightweight neural network can learn to represent both 3D point and line features.
In tests, our method secures a significant lead, marking the most considerable enhancement over state-of-the-art learning-based methodologies.
arXiv Detail & Related papers (2024-02-28T03:07:05Z) - ALSTER: A Local Spatio-Temporal Expert for Online 3D Semantic
Reconstruction [62.599588577671796]
We propose an online 3D semantic segmentation method that incrementally reconstructs a 3D semantic map from a stream of RGB-D frames.
Unlike offline methods, ours is directly applicable to scenarios with real-time constraints, such as robotics or mixed reality.
arXiv Detail & Related papers (2023-11-29T20:30:18Z) - PointOcc: Cylindrical Tri-Perspective View for Point-based 3D Semantic
Occupancy Prediction [72.75478398447396]
We propose a cylindrical tri-perspective view to represent point clouds effectively and comprehensively.
Considering the distance distribution of LiDAR point clouds, we construct the tri-perspective view in the cylindrical coordinate system.
We employ spatial group pooling to maintain structural details during projection and adopt 2D backbones to efficiently process each TPV plane.
arXiv Detail & Related papers (2023-08-31T17:57:17Z) - SeMLaPS: Real-time Semantic Mapping with Latent Prior Networks and
Quasi-Planar Segmentation [53.83313235792596]
We present a new methodology for real-time semantic mapping from RGB-D sequences.
It combines a 2D neural network and a 3D network based on a SLAM system with 3D occupancy mapping.
Our system achieves state-of-the-art semantic mapping quality within 2D-3D networks-based systems.
arXiv Detail & Related papers (2023-06-28T22:36:44Z) - IDLS: Inverse Depth Line based Visual-Inertial SLAM [9.38589798999922]
Inverse Depth Line SLAM (IDLS) is proposed to track the line features in SLAM in an accurate and efficient way.
IDLS is extensively evaluated in multiple perceptually-challenging datasets.
arXiv Detail & Related papers (2023-04-23T20:53:05Z) - 3D Line Mapping Revisited [86.13455066577657]
LIMAP is a library for 3D line mapping that robustly and efficiently creates 3D line maps from multi-view imagery.
Our code integrates seamlessly with existing point-based Structure-from-Motion methods.
Our robust 3D line maps also open up new research directions.
arXiv Detail & Related papers (2023-03-30T16:14:48Z) - LCPFormer: Towards Effective 3D Point Cloud Analysis via Local Context
Propagation in Transformers [60.51925353387151]
We propose a novel module named Local Context Propagation (LCP) to exploit the message passing between neighboring local regions.
We use the overlap points of adjacent local regions as intermediaries, then re-weight the features of these shared points from different local regions before passing them to the next layers.
The proposed method is applicable to different tasks and outperforms various transformer-based methods in benchmarks including 3D shape classification and dense prediction tasks.
arXiv Detail & Related papers (2022-10-23T15:43:01Z) - PL-VINS: Real-Time Monocular Visual-Inertial SLAM with Point and Line
Features [11.990163046319974]
This paper presents PL-VINS, a real-time optimization-based monocular VINS method with point and line features.
Experiments in a public benchmark dataset show that the localization error of our method is 12-16% less than that of VINS-Mono at the same pose update frequency.
arXiv Detail & Related papers (2020-09-16T04:27:33Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.