PAS-SLAM: A Visual SLAM System for Planar Ambiguous Scenes
- URL: http://arxiv.org/abs/2402.06131v1
- Date: Fri, 9 Feb 2024 01:34:26 GMT
- Title: PAS-SLAM: A Visual SLAM System for Planar Ambiguous Scenes
- Authors: Xinggang Hu, Yanmin Wu, Mingyuan Zhao, Linghao Yang, Xiangkui Zhang,
Xiangyang Ji
- Abstract summary: We propose a visual SLAM system based on planar features designed for planar ambiguous scenes.
We present an integrated data association strategy that combines plane parameters, semantic information, projection IoU, and non-parametric tests.
Finally, we design a set of multi-constraint factor graphs for camera pose optimization.
- Score: 41.47703182059505
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual SLAM (Simultaneous Localization and Mapping) based on planar features
has found widespread applications in fields such as environmental structure
perception and augmented reality. However, current research faces challenges in
accurately localizing and mapping in planar ambiguous scenes, primarily due to
the poor accuracy of the employed planar features and data association methods.
In this paper, we propose a visual SLAM system based on planar features
designed for planar ambiguous scenes, encompassing planar processing, data
association, and multi-constraint factor graph optimization. We introduce a
planar processing strategy that integrates semantic information with planar
features, extracting the edges and vertices of planes to be utilized in tasks
such as plane selection, data association, and pose optimization. Next, we
present an integrated data association strategy that combines plane parameters,
semantic information, projection IoU (Intersection over Union), and
non-parametric tests, achieving accurate and robust plane data association in
planar ambiguous scenes. Finally, we design a set of multi-constraint factor
graphs for camera pose optimization. Qualitative and quantitative experiments
conducted on publicly available datasets demonstrate that our proposed system
competes effectively in both accuracy and robustness in terms of map
construction and camera localization compared to state-of-the-art methods.
Related papers
- ActiveSplat: High-Fidelity Scene Reconstruction through Active Gaussian Splatting [12.628559736243536]
We propose ActiveSplat, an autonomous high-fidelity reconstruction system leveraging Gaussian splatting.
The system establishes a unified framework for online mapping, viewpoint selection, and path planning.
Experiments and ablation studies validate the efficacy of the proposed method in terms of reconstruction accuracy, data coverage, and exploration efficiency.
arXiv Detail & Related papers (2024-10-29T11:18:04Z) - Plane2Depth: Hierarchical Adaptive Plane Guidance for Monocular Depth Estimation [38.81275292687583]
We propose Plane2Depth, which adaptively utilizes plane information to improve depth prediction within a hierarchical framework.
In the proposed plane guided depth generator (PGDG), we design a set of plane queries as prototypes to softly model planes in the scene and predict per-pixel plane coefficients.
In the proposed adaptive plane query aggregation (APGA) module, we introduce a novel feature interaction approach to improve the aggregation of multi-scale plane features.
arXiv Detail & Related papers (2024-09-04T07:45:06Z) - Planning as In-Painting: A Diffusion-Based Embodied Task Planning
Framework for Environments under Uncertainty [56.30846158280031]
Task planning for embodied AI has been one of the most challenging problems.
We propose a task-agnostic method named 'planning as in-painting'
The proposed framework achieves promising performances in various embodied AI tasks.
arXiv Detail & Related papers (2023-12-02T10:07:17Z) - ShaRP: Shape-Regularized Multidimensional Projections [71.30697308446064]
We present a novel projection technique - ShaRP - that provides users explicit control over the visual signature of the created scatterplot.
ShaRP scales well with dimensionality and dataset size, and generically handles any quantitative dataset.
arXiv Detail & Related papers (2023-06-01T11:16:58Z) - PlanT: Explainable Planning Transformers via Object-Level
Representations [64.93938686101309]
PlanT is a novel approach for planning in the context of self-driving.
PlanT is based on imitation learning with a compact object-level input representation.
Our results indicate that PlanT can focus on the most relevant object in the scene, even when this object is geometrically distant.
arXiv Detail & Related papers (2022-10-25T17:59:46Z) - VIP-SLAM: An Efficient Tightly-Coupled RGB-D Visual Inertial Planar SLAM [25.681256050571058]
We propose a tightly-coupled SLAM system fused with RGB, Depth, IMU and structured plane information.
We use homography constraints to eliminate the parameters of numerous plane points in the optimization.
The global bundle adjustment is nearly 2 times faster than the sparse points based SLAM algorithm.
arXiv Detail & Related papers (2022-07-04T01:45:24Z) - Differentiable Spatial Planning using Transformers [87.90709874369192]
We propose Spatial Planning Transformers (SPT), which given an obstacle map learns to generate actions by planning over long-range spatial dependencies.
In the setting where the ground truth map is not known to the agent, we leverage pre-trained SPTs in an end-to-end framework.
SPTs outperform prior state-of-the-art differentiable planners across all the setups for both manipulation and navigation tasks.
arXiv Detail & Related papers (2021-12-02T06:48:16Z) - Visual SLAM with Graph-Cut Optimized Multi-Plane Reconstruction [11.215334675788952]
This paper presents a semantic planar SLAM system that improves pose estimation and mapping using cues from an instance planar segmentation network.
While the mainstream approaches are using RGB-D sensors, employing a monocular camera with such a system still faces challenges such as robust data association and precise geometric model fitting.
arXiv Detail & Related papers (2021-08-09T18:16:08Z) - Plan2Vec: Unsupervised Representation Learning by Latent Plans [106.37274654231659]
We introduce plan2vec, an unsupervised representation learning approach that is inspired by reinforcement learning.
Plan2vec constructs a weighted graph on an image dataset using near-neighbor distances, and then extrapolates this local metric to a global embedding by distilling path-integral over planned path.
We demonstrate the effectiveness of plan2vec on one simulated and two challenging real-world image datasets.
arXiv Detail & Related papers (2020-05-07T17:52:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.