VP-SLAM: A Monocular Real-time Visual SLAM with Points, Lines and
Vanishing Points
- URL: http://arxiv.org/abs/2210.12756v1
- Date: Sun, 23 Oct 2022 15:54:26 GMT
- Title: VP-SLAM: A Monocular Real-time Visual SLAM with Points, Lines and
Vanishing Points
- Authors: Andreas Georgis, Panagiotis Mermigkas, Petros Maragos
- Abstract summary: We present a real-time monocular Visual SLAM system that incorporates real-time methods for line and VP extraction.
We also present two strategies that exploit vanishing points to estimate the robot's translation and improve its rotation.
The proposed system achieves state-of-the-art results and runs in real time, and its performance remains close to the original ORB-SLAM2 system.
- Score: 31.55798962786664
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Traditional monocular Visual Simultaneous Localization and Mapping (vSLAM)
systems can be divided into three categories: those that use features, those
that rely on the image itself, and hybrid models. In the case of feature-based
methods, new research has evolved to incorporate more information from their
environment using geometric primitives beyond points, such as lines and planes.
This is because in many environments, which are man-made environments,
characterized as Manhattan world, geometric primitives such as lines and planes
occupy most of the space in the environment. The exploitation of these schemes
can lead to the introduction of algorithms capable of optimizing the trajectory
of a Visual SLAM system and also helping to construct an exuberant map. Thus,
we present a real-time monocular Visual SLAM system that incorporates real-time
methods for line and VP extraction, as well as two strategies that exploit
vanishing points to estimate the robot's translation and improve its
rotation.Particularly, we build on ORB-SLAM2, which is considered the current
state-of-the-art solution in terms of both accuracy and efficiency, and extend
its formulation to handle lines and VPs to create two strategies the first
optimize the rotation and the second refine the translation part from the known
rotation. First, we extract VPs using a real-time method and use them for a
global rotation optimization strategy. Second, we present a translation
estimation method that takes advantage of last-stage rotation optimization to
model a linear system. Finally, we evaluate our system on the TUM RGB-D
benchmark and demonstrate that the proposed system achieves state-of-the-art
results and runs in real time, and its performance remains close to the
original ORB-SLAM2 system
Related papers
- Double-Shot 3D Shape Measurement with a Dual-Branch Network [14.749887303860717]
We propose a dual-branch Convolutional Neural Network (CNN)-Transformer network (PDCNet) to process different structured light (SL) modalities.
Within PDCNet, a Transformer branch is used to capture global perception in the fringe images, while a CNN branch is designed to collect local details in the speckle images.
We show that our method can reduce fringe order ambiguity while producing high-accuracy results on a self-made dataset.
arXiv Detail & Related papers (2024-07-19T10:49:26Z) - An Accurate and Real-time Relative Pose Estimation from Triple Point-line Images by Decoupling Rotation and Translation [10.05584976985694]
3D-2D constraints provided by line features have been widely used in Visual Odometry (VO) and Structure-from-Motion (SfM) systems.
We propose a novel three-view pose solver based on rotation-translation decoupled estimation.
arXiv Detail & Related papers (2024-03-18T10:21:05Z) - DNS SLAM: Dense Neural Semantic-Informed SLAM [92.39687553022605]
DNS SLAM is a novel neural RGB-D semantic SLAM approach featuring a hybrid representation.
Our method integrates multi-view geometry constraints with image-based feature extraction to improve appearance details.
Our experimental results achieve state-of-the-art performance on both synthetic data and real-world data tracking.
arXiv Detail & Related papers (2023-11-30T21:34:44Z) - ALSTER: A Local Spatio-Temporal Expert for Online 3D Semantic
Reconstruction [62.599588577671796]
We propose an online 3D semantic segmentation method that incrementally reconstructs a 3D semantic map from a stream of RGB-D frames.
Unlike offline methods, ours is directly applicable to scenarios with real-time constraints, such as robotics or mixed reality.
arXiv Detail & Related papers (2023-11-29T20:30:18Z) - GS-SLAM: Dense Visual SLAM with 3D Gaussian Splatting [51.96353586773191]
We introduce textbfGS-SLAM that first utilizes 3D Gaussian representation in the Simultaneous Localization and Mapping system.
Our method utilizes a real-time differentiable splatting rendering pipeline that offers significant speedup to map optimization and RGB-D rendering.
Our method achieves competitive performance compared with existing state-of-the-art real-time methods on the Replica, TUM-RGBD datasets.
arXiv Detail & Related papers (2023-11-20T12:08:23Z) - RD-VIO: Robust Visual-Inertial Odometry for Mobile Augmented Reality in
Dynamic Environments [55.864869961717424]
It is typically challenging for visual or visual-inertial odometry systems to handle the problems of dynamic scenes and pure rotation.
We design a novel visual-inertial odometry (VIO) system called RD-VIO to handle both of these problems.
arXiv Detail & Related papers (2023-10-23T16:30:39Z) - Maximize to Explore: One Objective Function Fusing Estimation, Planning,
and Exploration [87.53543137162488]
We propose an easy-to-implement online reinforcement learning (online RL) framework called textttMEX.
textttMEX integrates estimation and planning components while balancing exploration exploitation automatically.
It can outperform baselines by a stable margin in various MuJoCo environments with sparse rewards.
arXiv Detail & Related papers (2023-05-29T17:25:26Z) - Structure PLP-SLAM: Efficient Sparse Mapping and Localization using
Point, Line and Plane for Monocular, RGB-D and Stereo Cameras [13.693353009049773]
This paper demonstrates a visual SLAM system that utilizes point and line cloud for robust camera localization, simultaneously, with an embedded piece-wise planar reconstruction (PPR) module.
We address the challenge of reconstructing geometric primitives with scale ambiguity by proposing several run-time optimizations on the reconstructed lines and planes.
The results show that our proposed SLAM tightly incorporates the semantic features to boost both tracking as well as backend optimization.
arXiv Detail & Related papers (2022-07-13T09:05:35Z) - Keeping Less is More: Point Sparsification for Visual SLAM [1.370633147306388]
This study proposes an efficient graph optimization for sparsifying map points in SLAM systems.
Specifically, we formulate a maximum pose-visibility and maximum spatial diversity problem as a minimum-cost maximum-flow graph optimization problem.
The proposed method works as an additional step in existing SLAM systems, so it can be used in both conventional or learning based SLAM systems.
arXiv Detail & Related papers (2022-07-01T06:39:38Z) - Greedy-Based Feature Selection for Efficient LiDAR SLAM [12.257338124961622]
This paper demonstrates that actively selecting a subset of features significantly improves both the accuracy and efficiency of an L-SLAM system.
We show that our approach exhibits low localization error and speedup compared to the state-of-the-art L-SLAM systems.
arXiv Detail & Related papers (2021-03-24T11:03:16Z) - Pushing the Envelope of Rotation Averaging for Visual SLAM [69.7375052440794]
We propose a novel optimization backbone for visual SLAM systems.
We leverage averaging to improve the accuracy, efficiency and robustness of conventional monocular SLAM systems.
Our approach can exhibit up to 10x faster with comparable accuracy against the state-art on public benchmarks.
arXiv Detail & Related papers (2020-11-02T18:02:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.