360ORB-SLAM: A Visual SLAM System for Panoramic Images with Depth
Completion Network
- URL: http://arxiv.org/abs/2401.10560v1
- Date: Fri, 19 Jan 2024 08:52:24 GMT
- Title: 360ORB-SLAM: A Visual SLAM System for Panoramic Images with Depth
Completion Network
- Authors: Yichen Chen, Yiqi Pan, Ruyu Liu, Haoyu Zhang, Guodao Zhang, Bo Sun and
Jianhua Zhang
- Abstract summary: This paper proposes a 360ORB-SLAM system for panoramic images that combines with a depth completion network.
The proposed method achieves superior scale accuracy compared to existing monocular SLAM methods.
The integration of the depth completion network enhances system stability and mitigates the impact of dynamic elements on SLAM performance.
- Score: 18.23570356507258
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To enhance the performance and effect of AR/VR applications and visual
assistance and inspection systems, visual simultaneous localization and mapping
(vSLAM) is a fundamental task in computer vision and robotics. However,
traditional vSLAM systems are limited by the camera's narrow field-of-view,
resulting in challenges such as sparse feature distribution and lack of dense
depth information. To overcome these limitations, this paper proposes a
360ORB-SLAM system for panoramic images that combines with a depth completion
network. The system extracts feature points from the panoramic image, utilizes
a panoramic triangulation module to generate sparse depth information, and
employs a depth completion network to obtain a dense panoramic depth map.
Experimental results on our novel panoramic dataset constructed based on Carla
demonstrate that the proposed method achieves superior scale accuracy compared
to existing monocular SLAM methods and effectively addresses the challenges of
feature association and scale ambiguity. The integration of the depth
completion network enhances system stability and mitigates the impact of
dynamic elements on SLAM performance.
Related papers
- PVAFN: Point-Voxel Attention Fusion Network with Multi-Pooling Enhancing for 3D Object Detection [59.355022416218624]
integration of point and voxel representations is becoming more common in LiDAR-based 3D object detection.
We propose a novel two-stage 3D object detector, called Point-Voxel Attention Fusion Network (PVAFN)
PVAFN uses a multi-pooling strategy to integrate both multi-scale and region-specific information effectively.
arXiv Detail & Related papers (2024-08-26T19:43:01Z) - MoD-SLAM: Monocular Dense Mapping for Unbounded 3D Scene Reconstruction [2.3630527334737104]
MoD-SLAM is the first monocular NeRF-based dense mapping method that allows 3D reconstruction in real-time in unbounded scenes.
By introducing a robust depth loss term into the tracking process, our SLAM system achieves more precise pose estimation in large-scale scenes.
Our experiments on two standard datasets show that MoD-SLAM achieves competitive performance, improving the accuracy of the 3D reconstruction and localization by up to 30% and 15% respectively.
arXiv Detail & Related papers (2024-02-06T07:07:33Z) - Depth Completion with Multiple Balanced Bases and Confidence for Dense
Monocular SLAM [34.78726455243436]
We propose a novel method that integrates a light-weight depth completion network into a sparse SLAM system.
Specifically, we present a specifically optimized multi-basis depth completion network, called BBC-Net.
BBC-Net can predict multiple balanced bases and a confidence map from a monocular image with sparse points generated by off-the-shelf keypoint-based SLAM systems.
arXiv Detail & Related papers (2023-09-08T06:15:27Z) - HDPV-SLAM: Hybrid Depth-augmented Panoramic Visual SLAM for Mobile
Mapping System with Tilted LiDAR and Panoramic Visual Camera [4.2421412410466575]
This paper proposes a novel visual simultaneous localization and mapping (SLAM) system called Hybrid Depth-augmented Panoramic Visual SLAM (HDPV-SLAM)
It employs a panoramic camera and a tilted multi-beam LiDAR scanner to generate accurate and metrically-scaled trajectories.
It aims to solve the two major issues hindering the performance of similar SLAM systems.
arXiv Detail & Related papers (2023-01-27T16:25:28Z) - TANDEM: Tracking and Dense Mapping in Real-time using Deep Multi-view
Stereo [55.30992853477754]
We present TANDEM, a real-time monocular tracking and dense framework.
For pose estimation, TANDEM performs photometric bundle adjustment based on a sliding window of alignments.
TANDEM shows state-of-the-art real-time 3D reconstruction performance.
arXiv Detail & Related papers (2021-11-14T19:01:02Z) - SLIDE: Single Image 3D Photography with Soft Layering and Depth-aware
Inpainting [54.419266357283966]
Single image 3D photography enables viewers to view a still image from novel viewpoints.
Recent approaches combine monocular depth networks with inpainting networks to achieve compelling results.
We present SLIDE, a modular and unified system for single image 3D photography.
arXiv Detail & Related papers (2021-09-02T16:37:20Z) - DSP-SLAM: Object Oriented SLAM with Deep Shape Priors [16.867669408751507]
We propose an object-oriented SLAM system that builds a rich and accurate joint map of dense 3D models for foreground objects.
DSP-SLAM takes as input the 3D point cloud reconstructed by a feature-based SLAM system.
Our evaluation shows improvements in object pose and shape reconstruction with respect to recent deep prior-based reconstruction methods.
arXiv Detail & Related papers (2021-08-21T10:00:12Z) - Light Field Reconstruction Using Convolutional Network on EPI and
Extended Applications [78.63280020581662]
A novel convolutional neural network (CNN)-based framework is developed for light field reconstruction from a sparse set of views.
We demonstrate the high performance and robustness of the proposed framework compared with state-of-the-art algorithms.
arXiv Detail & Related papers (2021-03-24T08:16:32Z) - Pseudo RGB-D for Self-Improving Monocular SLAM and Depth Prediction [72.30870535815258]
CNNs for monocular depth prediction represent two largely disjoint approaches towards building a 3D map of the surrounding environment.
We propose a joint narrow and wide baseline based self-improving framework, where on the one hand the CNN-predicted depth is leveraged to perform pseudo RGB-D feature-based SLAM.
On the other hand, the bundle-adjusted 3D scene structures and camera poses from the more principled geometric SLAM are injected back into the depth network through novel wide baseline losses.
arXiv Detail & Related papers (2020-04-22T16:31:59Z) - OmniSLAM: Omnidirectional Localization and Dense Mapping for
Wide-baseline Multi-camera Systems [88.41004332322788]
We present an omnidirectional localization and dense mapping system for a wide-baseline multiview stereo setup with ultra-wide field-of-view (FOV) fisheye cameras.
For more practical and accurate reconstruction, we first introduce improved and light-weighted deep neural networks for the omnidirectional depth estimation.
We integrate our omnidirectional depth estimates into the visual odometry (VO) and add a loop closing module for global consistency.
arXiv Detail & Related papers (2020-03-18T05:52:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.