AQUA-SLAM: Tightly-Coupled Underwater Acoustic-Visual-Inertial SLAM with Sensor Calibration
- URL: http://arxiv.org/abs/2503.11420v1
- Date: Fri, 14 Mar 2025 14:03:33 GMT
- Title: AQUA-SLAM: Tightly-Coupled Underwater Acoustic-Visual-Inertial SLAM with Sensor Calibration
- Authors: Shida Xu, Kaicheng Zhang, Sen Wang,
- Abstract summary: This paper introduces a novel, tightly-coupled Acoustic-Visual-Inertial SLAM approach.<n>It fuses a Doppler Velocity Log (DVL), a stereo camera, and an Inertial Measurement Unit (IMU) within a graph optimization framework.<n>We show that our method surpasses current state-of-the-art underwater and visual-inertial SLAM systems in terms of localization accuracy and robustness.
- Score: 9.727021664098485
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Underwater environments pose significant challenges for visual Simultaneous Localization and Mapping (SLAM) systems due to limited visibility, inadequate illumination, and sporadic loss of structural features in images. Addressing these challenges, this paper introduces a novel, tightly-coupled Acoustic-Visual-Inertial SLAM approach, termed AQUA-SLAM, to fuse a Doppler Velocity Log (DVL), a stereo camera, and an Inertial Measurement Unit (IMU) within a graph optimization framework. Moreover, we propose an efficient sensor calibration technique, encompassing multi-sensor extrinsic calibration (among the DVL, camera and IMU) and DVL transducer misalignment calibration, with a fast linear approximation procedure for real-time online execution. The proposed methods are extensively evaluated in a tank environment with ground truth, and validated for offshore applications in the North Sea. The results demonstrate that our method surpasses current state-of-the-art underwater and visual-inertial SLAM systems in terms of localization accuracy and robustness. The proposed system will be made open-source for the community.
Related papers
- Spike Imaging Velocimetry: Dense Motion Estimation of Fluids Using Spike Cameras [56.30609439239127]
This study investigates the tremendous potential of spike cameras (a type of ultra-high-speed, high-dynamic-range camera) in Particle Image Velocimetry (PIV)
We propose a deep learning framework, Spike Imaging Velocimetry (SIV), designed specifically for highly turbulent and intricate flow fields.
We present a spike-based PIV dataset, Particle Scenes with Spike and Displacement (PSSD), which provides labeled data for three challenging fluid dynamics scenarios.
arXiv Detail & Related papers (2025-04-26T09:14:45Z) - A Data-Driven Method for INS/DVL Alignment [2.915868985330569]
Inertial sensors and Doppler velocity logs (DVL) fusion is a promising solution for long-range underwater navigation.
We propose an end-to-end deep learning framework for the alignment process.
arXiv Detail & Related papers (2025-03-27T10:38:37Z) - KFD-NeRF: Rethinking Dynamic NeRF with Kalman Filter [49.85369344101118]
We introduce KFD-NeRF, a novel dynamic neural radiance field integrated with an efficient and high-quality motion reconstruction framework based on Kalman filtering.
Our key idea is to model the dynamic radiance field as a dynamic system whose temporally varying states are estimated based on two sources of knowledge: observations and predictions.
Our KFD-NeRF demonstrates similar or even superior performance within comparable computational time and state-of-the-art view synthesis performance with thorough training.
arXiv Detail & Related papers (2024-07-18T05:48:24Z) - Attenuation-Aware Weighted Optical Flow with Medium Transmission Map for Learning-based Visual Odometry in Underwater terrain [0.03749861135832072]
This paper addresses the challenge of improving learning-based monocular visual odometry (VO) in underwater environments.
The novel wflow-TartanVO is introduced, enhancing the accuracy of VO systems for autonomous underwater vehicles (AUVs)
Evaluation of different real-world underwater datasets demonstrates the outperformance of wflow-TartanVO over baseline VO methods.
arXiv Detail & Related papers (2024-07-18T05:00:15Z) - SLAIM: Robust Dense Neural SLAM for Online Tracking and Mapping [15.63276368052395]
We propose a novel coarse-to-fine tracking model tailored for Neural Radiance Field SLAM (NeRF-SLAM)
Existing NeRF-SLAM systems consistently exhibit inferior tracking performance compared to traditional SLAM algorithms.
We implement both local and global bundle-adjustment to produce a robust (coarse-to-fine) and accurate (KL regularizer) SLAM solution.
arXiv Detail & Related papers (2024-04-17T14:23:28Z) - AscDAMs: Advanced SLAM-based channel detection and mapping system [10.736648466109045]
We propose an advanced SLAM-based channel detection and mapping system, namely AscDAMs.
It features three main enhancements to post-process SLAM results.
Two field experiments were conducted in Chutou Gully, Wenchuan County in China in February and November 2023.
arXiv Detail & Related papers (2024-01-25T01:22:29Z) - GAFlow: Incorporating Gaussian Attention into Optical Flow [62.646389181507764]
We push Gaussian Attention (GA) into the optical flow models to accentuate local properties during representation learning.
We introduce a novel Gaussian-Constrained Layer (GCL) which can be easily plugged into existing Transformer blocks.
For reliable motion analysis, we provide a new Gaussian-Guided Attention Module (GGAM)
arXiv Detail & Related papers (2023-09-28T07:46:01Z) - AMSP-UOD: When Vortex Convolution and Stochastic Perturbation Meet
Underwater Object Detection [40.532331552038485]
We present a novel Amplitude-Modulated Perturbation and Vortex Convolutional Network, AMSP-UOD.
AMSP-UOD addresses the impact of non-ideal imaging factors on detection accuracy in complex underwater environments.
Our method outperforms existing state-of-the-art methods in terms of accuracy and noise immunity.
arXiv Detail & Related papers (2023-08-23T05:03:45Z) - Towards Scale-Aware, Robust, and Generalizable Unsupervised Monocular
Depth Estimation by Integrating IMU Motion Dynamics [74.1720528573331]
Unsupervised monocular depth and ego-motion estimation has drawn extensive research attention in recent years.
We propose DynaDepth, a novel scale-aware framework that integrates information from vision and IMU motion dynamics.
We validate the effectiveness of DynaDepth by conducting extensive experiments and simulations on the KITTI and Make3D datasets.
arXiv Detail & Related papers (2022-07-11T07:50:22Z) - Sensor-Guided Optical Flow [53.295332513139925]
This paper proposes a framework to guide an optical flow network with external cues to achieve superior accuracy on known or unseen domains.
We show how these can be obtained by combining depth measurements from active sensors with geometry and hand-crafted optical flow algorithms.
arXiv Detail & Related papers (2021-09-30T17:59:57Z) - Pushing the Envelope of Rotation Averaging for Visual SLAM [69.7375052440794]
We propose a novel optimization backbone for visual SLAM systems.
We leverage averaging to improve the accuracy, efficiency and robustness of conventional monocular SLAM systems.
Our approach can exhibit up to 10x faster with comparable accuracy against the state-art on public benchmarks.
arXiv Detail & Related papers (2020-11-02T18:02:26Z) - Joint Unsupervised Learning of Optical Flow and Egomotion with Bi-Level
Optimization [59.9673626329892]
We exploit the global relationship between optical flow and camera motion using epipolar geometry.
We use implicit differentiation to enable back-propagation through the lower-level geometric optimization layer independent of its implementation.
arXiv Detail & Related papers (2020-02-26T22:28:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.