DROID-SLAM: Deep Visual SLAM for Monocular, Stereo, and RGB-D Cameras
- URL: http://arxiv.org/abs/2108.10869v1
- Date: Tue, 24 Aug 2021 17:50:10 GMT
- Title: DROID-SLAM: Deep Visual SLAM for Monocular, Stereo, and RGB-D Cameras
- Authors: Zachary Teed and Jia Deng
- Abstract summary: DROID-SLAM is a new deep learning based SLAM system.
It can leverage stereo or RGB-D video to achieve improved performance at test time.
- Score: 71.41252518419486
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce DROID-SLAM, a new deep learning based SLAM system. DROID-SLAM
consists of recurrent iterative updates of camera pose and pixelwise depth
through a Dense Bundle Adjustment layer. DROID-SLAM is accurate, achieving
large improvements over prior work, and robust, suffering from substantially
fewer catastrophic failures. Despite training on monocular video, it can
leverage stereo or RGB-D video to achieve improved performance at test time.
The URL to our open source code is https://github.com/princeton-vl/DROID-SLAM.
Related papers
- Dynamic EventNeRF: Reconstructing General Dynamic Scenes from Multi-view RGB and Event Streams [69.65147723239153]
Volumetric reconstruction of dynamic scenes is an important problem in computer vision.
It is especially challenging in poor lighting and with fast motion.
We propose the first method totemporally reconstruct a scene from sparse multi-view event streams and sparse RGB frames.
arXiv Detail & Related papers (2024-12-09T18:56:18Z) - FlashSLAM: Accelerated RGB-D SLAM for Real-Time 3D Scene Reconstruction with Gaussian Splatting [14.130327598928778]
FlashSLAM is a novel SLAM approach that leverages 3D Gaussian Splatting for efficient and robust 3D scene reconstruction.
Existing 3DGS-based SLAM methods often fall short in sparse view settings and during large camera movements.
Our method achieves up to a 92% improvement in average tracking accuracy over previous methods.
arXiv Detail & Related papers (2024-12-01T05:44:38Z) - DROID-Splat: Combining end-to-end SLAM with 3D Gaussian Splatting [3.2771631221674333]
We introduce a SLAM system based on an end-to-end Tracker and extend it with a Renderer based on recent 3D Gaussian Splatting techniques.
Our framework textbfDroidSplat achieves both SotA tracking and rendering results on common SLAM benchmarks.
arXiv Detail & Related papers (2024-11-26T18:25:51Z) - EvenNICER-SLAM: Event-based Neural Implicit Encoding SLAM [69.83383687049994]
We propose EvenNICER-SLAM, a novel approach to dense visual simultaneous localization and mapping.
EvenNICER-SLAM incorporates event cameras that respond to intensity changes instead of absolute brightness.
Our results suggest the potential for event cameras to improve the robustness of dense SLAM systems against fast camera motion in real-world scenarios.
arXiv Detail & Related papers (2024-10-04T13:52:01Z) - Deep Patch Visual SLAM [29.81779535397074]
We introduce Deep Patch Visual (DPV) SLAM, a method for monocular visual SLAM on a single GPU.
DPV-SLAM maintains a high minimum framerate and small memory overhead (5-7G) compared to existing deep SLAM systems.
We achieve comparable accuracy to DROID-SLAM on EuRoC and TartanAir while running 2.5x faster using a fraction of the memory.
arXiv Detail & Related papers (2024-08-03T03:51:47Z) - HDR-GS: Efficient High Dynamic Range Novel View Synthesis at 1000x Speed via Gaussian Splatting [76.5908492298286]
Existing HDR NVS methods are mainly based on NeRF.
They suffer from long training time and slow inference speed.
We propose a new framework, High Dynamic Range Gaussian Splatting (-GS)
arXiv Detail & Related papers (2024-05-24T00:46:58Z) - GlORIE-SLAM: Globally Optimized RGB-only Implicit Encoding Point Cloud SLAM [53.6402869027093]
We propose an efficient RGB-only dense SLAM system using a flexible neural point cloud representation scene.
We also introduce a novel DSPO layer for bundle adjustment which optimize the pose and depth of implicits along with the scale of the monocular depth.
arXiv Detail & Related papers (2024-03-28T16:32:06Z) - Volume-DROID: A Real-Time Implementation of Volumetric Mapping with
DROID-SLAM [0.0]
Volume-DROID is a novel approach for Simultaneous localization and Mapping (SLAM)
It combines DROID-SLAM, point cloud registration, an off-the-shelf semantic segmentation network, and Convolutional Bayesian Kernel Inference (ConvBKI)
The key innovation of our method is the real-time fusion of DROID-SLAM and Convolutional Bayesian Kernel Inference (ConvBKI)
arXiv Detail & Related papers (2023-06-12T03:50:50Z) - Seeing Through The Noisy Dark: Toward Real-world Low-Light Image
Enhancement and Denoising [125.56062454927755]
Real-world low-light environment usually suffer from lower visibility and heavier noise, due to insufficient light or hardware limitation.
We propose a novel end-to-end method termed Real-world Low-light Enhancement & Denoising Network (RLED-Net)
arXiv Detail & Related papers (2022-10-02T14:57:23Z) - Deep Parametric 3D Filters for Joint Video Denoising and Illumination
Enhancement in Video Super Resolution [96.89588203312451]
This paper presents a new parametric representation called Deep Parametric 3D Filters (DP3DF)
DP3DF incorporates local information to enable simultaneous denoising, illumination enhancement, and SR efficiently in a single encoder-and-decoder network.
Also, a dynamic residual frame is jointly learned with the DP3DF via a shared backbone to further boost the SR quality.
arXiv Detail & Related papers (2022-07-05T03:57:25Z) - Differentiable SLAM-net: Learning Particle SLAM for Visual Navigation [15.677860200178959]
SLAM-net encodes a particle filter based SLAM algorithm in a differentiable graph.
It learns task-oriented neural network components by backpropagating through the SLAM algorithm.
It significantly outperforms the widely adapted ORB-SLAM in noisy conditions.
arXiv Detail & Related papers (2021-05-17T03:54:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.