Tightly-Coupled LiDAR-Visual SLAM Based on Geometric Features for Mobile
Agents
- URL: http://arxiv.org/abs/2307.07763v3
- Date: Tue, 26 Dec 2023 04:26:32 GMT
- Title: Tightly-Coupled LiDAR-Visual SLAM Based on Geometric Features for Mobile
Agents
- Authors: Ke Cao, Ruiping Liu, Ze Wang, Kunyu Peng, Jiaming Zhang, Junwei Zheng,
Zhifeng Teng, Kailun Yang, Rainer Stiefelhagen
- Abstract summary: We propose a tightly-coupled LiDAR-visual SLAM based on geometric features.
The entire line segment detected by the visual subsystem overcomes the limitation of the LiDAR subsystem.
Our system achieves more accurate and robust pose estimation compared to current state-of-the-art multi-modal methods.
- Score: 43.137917788594926
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The mobile robot relies on SLAM (Simultaneous Localization and Mapping) to
provide autonomous navigation and task execution in complex and unknown
environments. However, it is hard to develop a dedicated algorithm for mobile
robots due to dynamic and challenging situations, such as poor lighting
conditions and motion blur. To tackle this issue, we propose a tightly-coupled
LiDAR-visual SLAM based on geometric features, which includes two sub-systems
(LiDAR and monocular visual SLAM) and a fusion framework. The fusion framework
associates the depth and semantics of the multi-modal geometric features to
complement the visual line landmarks and to add direction optimization in
Bundle Adjustment (BA). This further constrains visual odometry. On the other
hand, the entire line segment detected by the visual subsystem overcomes the
limitation of the LiDAR subsystem, which can only perform the local calculation
for geometric features. It adjusts the direction of linear feature points and
filters out outliers, leading to a higher accurate odometry system. Finally, we
employ a module to detect the subsystem's operation, providing the LiDAR
subsystem's output as a complementary trajectory to our system while visual
subsystem tracking fails. The evaluation results on the public dataset M2DGR,
gathered from ground robots across various indoor and outdoor scenarios, show
that our system achieves more accurate and robust pose estimation compared to
current state-of-the-art multi-modal methods.
Related papers
- LiDAR-GS:Real-time LiDAR Re-Simulation using Gaussian Splatting [50.808933338389686]
LiDAR simulation plays a crucial role in closed-loop simulation for autonomous driving.
We present LiDAR-GS, the first LiDAR Gaussian Splatting method, for real-time high-fidelity re-simulation of LiDAR sensor scans in public urban road scenes.
Our approach succeeds in simultaneously re-simulating depth, intensity, and ray-drop channels, achieving state-of-the-art results in both rendering frame rate and quality on publically available large scene datasets.
arXiv Detail & Related papers (2024-10-07T15:07:56Z) - FAST-LIVO2: Fast, Direct LiDAR-Inertial-Visual Odometry [28.606325312582218]
We propose FAST-LIVO2, a fast, direct LiDAR-inertial-visual odometry framework to achieve accurate and robust state estimation in SLAM tasks.
FAST-LIVO2 fuses the IMU, LiDAR and image measurements efficiently through a sequential update strategy.
We show three applications of FAST-LIVO2, including real-time onboard navigation, airborne mapping, and 3D model rendering.
arXiv Detail & Related papers (2024-08-26T06:01:54Z) - RD-VIO: Robust Visual-Inertial Odometry for Mobile Augmented Reality in
Dynamic Environments [55.864869961717424]
It is typically challenging for visual or visual-inertial odometry systems to handle the problems of dynamic scenes and pure rotation.
We design a novel visual-inertial odometry (VIO) system called RD-VIO to handle both of these problems.
arXiv Detail & Related papers (2023-10-23T16:30:39Z) - Structure PLP-SLAM: Efficient Sparse Mapping and Localization using
Point, Line and Plane for Monocular, RGB-D and Stereo Cameras [13.693353009049773]
This paper demonstrates a visual SLAM system that utilizes point and line cloud for robust camera localization, simultaneously, with an embedded piece-wise planar reconstruction (PPR) module.
We address the challenge of reconstructing geometric primitives with scale ambiguity by proposing several run-time optimizations on the reconstructed lines and planes.
The results show that our proposed SLAM tightly incorporates the semantic features to boost both tracking as well as backend optimization.
arXiv Detail & Related papers (2022-07-13T09:05:35Z) - Multitask AET with Orthogonal Tangent Regularity for Dark Object
Detection [84.52197307286681]
We propose a novel multitask auto encoding transformation (MAET) model to enhance object detection in a dark environment.
In a self-supervision manner, the MAET learns the intrinsic visual structure by encoding and decoding the realistic illumination-degrading transformation.
We have achieved the state-of-the-art performance using synthetic and real-world datasets.
arXiv Detail & Related papers (2022-05-06T16:27:14Z) - An Online Semantic Mapping System for Extending and Enhancing Visual
SLAM [2.538209532048867]
We present a real-time semantic mapping approach for mobile vision systems with a 2D to 3D object detection pipeline and rapid data association for generated landmarks.
Our system reaches real-time capabilities with an average iteration duration of 65ms and is able to improve the pose estimation of a state-of-the-art SLAM by up to 68% on a public dataset.
arXiv Detail & Related papers (2022-03-08T09:14:37Z) - Visual SLAM with Graph-Cut Optimized Multi-Plane Reconstruction [11.215334675788952]
This paper presents a semantic planar SLAM system that improves pose estimation and mapping using cues from an instance planar segmentation network.
While the mainstream approaches are using RGB-D sensors, employing a monocular camera with such a system still faces challenges such as robust data association and precise geometric model fitting.
arXiv Detail & Related papers (2021-08-09T18:16:08Z) - Pushing the Envelope of Rotation Averaging for Visual SLAM [69.7375052440794]
We propose a novel optimization backbone for visual SLAM systems.
We leverage averaging to improve the accuracy, efficiency and robustness of conventional monocular SLAM systems.
Our approach can exhibit up to 10x faster with comparable accuracy against the state-art on public benchmarks.
arXiv Detail & Related papers (2020-11-02T18:02:26Z) - Robust Odometry and Mapping for Multi-LiDAR Systems with Online
Extrinsic Calibration [15.946728828122385]
This paper proposes a system to achieve robust and simultaneous extrinsic calibration, odometry, and mapping for multiple LiDARs.
We validate our approach's performance with extensive experiments on ten sequences (4.60km total length) for the calibration and SLAM.
We demonstrate that the proposed work is a complete, robust, and system for various multi-LiDAR setups.
arXiv Detail & Related papers (2020-10-27T13:51:26Z) - Risk-Averse MPC via Visual-Inertial Input and Recurrent Networks for
Online Collision Avoidance [95.86944752753564]
We propose an online path planning architecture that extends the model predictive control (MPC) formulation to consider future location uncertainties.
Our algorithm combines an object detection pipeline with a recurrent neural network (RNN) which infers the covariance of state estimates.
The robustness of our methods is validated on complex quadruped robot dynamics and can be generally applied to most robotic platforms.
arXiv Detail & Related papers (2020-07-28T07:34:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.