Deep Patch Visual SLAM
- URL: http://arxiv.org/abs/2408.01654v1
- Date: Sat, 3 Aug 2024 03:51:47 GMT
- Title: Deep Patch Visual SLAM
- Authors: Lahav Lipson, Zachary Teed, Jia Deng,
- Abstract summary: We introduce Deep Patch Visual (DPV) SLAM, a method for monocular visual SLAM on a single GPU.
DPV-SLAM maintains a high minimum framerate and small memory overhead (5-7G) compared to existing deep SLAM systems.
We achieve comparable accuracy to DROID-SLAM on EuRoC and TartanAir while running 2.5x faster using a fraction of the memory.
- Score: 29.81779535397074
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work in visual SLAM has shown the effectiveness of using deep network backbones. Despite excellent accuracy, however, such approaches are often expensive to run or do not generalize well zero-shot. Their runtime can also fluctuate wildly while their frontend and backend fight for access to GPU resources. To address these problems, we introduce Deep Patch Visual (DPV) SLAM, a method for monocular visual SLAM on a single GPU. DPV-SLAM maintains a high minimum framerate and small memory overhead (5-7G) compared to existing deep SLAM systems. On real-world datasets, DPV-SLAM runs at 1x-4x real-time framerates. We achieve comparable accuracy to DROID-SLAM on EuRoC and TartanAir while running 2.5x faster using a fraction of the memory. DPV-SLAM is an extension to the DPVO visual odometry system; its code can be found in the same repository: https://github.com/princeton-vl/DPVO
Related papers
- DROID-Splat: Combining end-to-end SLAM with 3D Gaussian Splatting [3.2771631221674333]
We introduce a SLAM system based on an end-to-end Tracker and extend it with a Renderer based on recent 3D Gaussian Splatting techniques.
Our framework textbfDroidSplat achieves both SotA tracking and rendering results on common SLAM benchmarks.
arXiv Detail & Related papers (2024-11-26T18:25:51Z) - High-Speed Stereo Visual SLAM for Low-Powered Computing Devices [35.76305042835835]
We present an accurate and GPU-accelerated Stereo Visual SLAM design called Jetson-SLAM.
It exhibits frame-processing rates above 60FPS on NVIDIA's low-powered 10W Jetson-NX embedded computer.
arXiv Detail & Related papers (2024-10-05T09:16:44Z) - EvenNICER-SLAM: Event-based Neural Implicit Encoding SLAM [69.83383687049994]
We propose EvenNICER-SLAM, a novel approach to dense visual simultaneous localization and mapping.
EvenNICER-SLAM incorporates event cameras that respond to intensity changes instead of absolute brightness.
Our results suggest the potential for event cameras to improve the robustness of dense SLAM systems against fast camera motion in real-world scenarios.
arXiv Detail & Related papers (2024-10-04T13:52:01Z) - Loopy-SLAM: Dense Neural SLAM with Loop Closures [53.11936461015725]
We introduce Loopy-SLAM that globally optimize poses and the dense 3D model.
We use frame-to-model tracking using a data-driven point-based submap generation method and trigger loop closures online by performing global place recognition.
Evaluation on the synthetic Replica and real-world TUM-RGBD and ScanNet datasets demonstrate competitive or superior performance in tracking, mapping, and rendering accuracy when compared to existing dense neural RGBD SLAM methods.
arXiv Detail & Related papers (2024-02-14T18:18:32Z) - Orbeez-SLAM: A Real-time Monocular Visual SLAM with ORB Features and
NeRF-realized Mapping [18.083667773491083]
We develop a visual SLAM that adapts to new scenes without pre-training and generates dense maps for downstream tasks in real-time.
Orbeez-SLAM collaborates with implicit neural representation (NeRF) and visual odometry to achieve our goals.
Results show that our SLAM is up to 800x faster than the strong baseline with superior rendering outcomes.
arXiv Detail & Related papers (2022-09-27T09:37:57Z) - Deep Patch Visual Odometry [66.8086971254714]
Deep Patch Visual Odometry (DPVO) is a new deep learning system for monocular Visual Odometry (VO)
DPVO uses a novel recurrent network architecture designed for tracking image patches across time.
On Standard benchmarks, DPVO outperforms all prior work, including the learning-based state-of-the-art VO-system.
arXiv Detail & Related papers (2022-08-08T04:47:38Z) - DROID-SLAM: Deep Visual SLAM for Monocular, Stereo, and RGB-D Cameras [71.41252518419486]
DROID-SLAM is a new deep learning based SLAM system.
It can leverage stereo or RGB-D video to achieve improved performance at test time.
arXiv Detail & Related papers (2021-08-24T17:50:10Z) - OV$^{2}$SLAM : A Fully Online and Versatile Visual SLAM for Real-Time
Applications [59.013743002557646]
We describe OV$2$SLAM, a fully online algorithm, handling both monocular and stereo camera setups, various map scales and frame-rates ranging from a few Hertz up to several hundreds.
For the benefit of the community, we release the source code: urlhttps://github.com/ov2slam/ov2slam.
arXiv Detail & Related papers (2021-02-08T08:39:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.