FeatureSLAM: Feature-enriched 3D gaussian splatting SLAM in real time
- URL: http://arxiv.org/abs/2601.05738v1
- Date: Fri, 09 Jan 2026 11:40:16 GMT
- Title: FeatureSLAM: Feature-enriched 3D gaussian splatting SLAM in real time
- Authors: Christopher Thirgood, Oscar Mendez, Erin Ling, Jon Storey, Simon Hadfield,
- Abstract summary: We present a real-time tracking SLAM system that unifies efficient camera tracking and feature-enriched mapping using 3D Gaussian Splatting (3DGS)<n>Our main contribution is integrating dense feature semanticization with semanticization into the novel-view synthesis aligned with a visual foundation model.<n>We obtain 9% lower pose error and 8% higher mapping accuracy compared to recent fixed-set SLAM baselines.
- Score: 11.883404434697809
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a real-time tracking SLAM system that unifies efficient camera tracking with photorealistic feature-enriched mapping using 3D Gaussian Splatting (3DGS). Our main contribution is integrating dense feature rasterization into the novel-view synthesis, aligned with a visual foundation model. This yields strong semantics, going beyond basic RGB-D input, aiding both tracking and mapping accuracy. Unlike previous semantic SLAM approaches (which embed pre-defined class labels) FeatureSLAM enables entirely new downstream tasks via free-viewpoint, open-set segmentation. Across standard benchmarks, our method achieves real-time tracking, on par with state-of-the-art systems while improving tracking stability and map fidelity without prohibitive compute. Quantitatively, we obtain 9\% lower pose error and 8\% higher mapping accuracy compared to recent fixed-set SLAM baselines. Our results confirm that real-time feature-embedded SLAM, is not only valuable for enabling new downstream applications. It also improves the performance of the underlying tracking and mapping subsystems, providing semantic and language masking results that are on-par with offline 3DGS models, alongside state-of-the-art tracking, depth and RGB rendering.
Related papers
- VIGS-SLAM: Visual Inertial Gaussian Splatting SLAM [75.55522219717137]
We present VIGS-SLAM, a visual-inertial 3D Gaussian Splatting SLAM system.<n>It achieves robust real-time tracking and high-fidelity reconstruction.<n>Our method tightly couples visual and inertial cues within a unified optimization framework.
arXiv Detail & Related papers (2025-12-02T00:19:13Z) - SING3R-SLAM: Submap-based Indoor Monocular Gaussian SLAM with 3D Reconstruction Priors [80.51557267896938]
SING3R-SLAM is a globally consistent and compact Gaussian-based dense RGB SLAM framework.<n>We show that SING3R-SLAM achieves state-of-the-art tracking, 3D reconstruction, and novel view rendering, resulting in over 12% improvement in tracking and producing finer, more detailed geometry.
arXiv Detail & Related papers (2025-11-21T12:40:55Z) - Pseudo Depth Meets Gaussian: A Feed-forward RGB SLAM Baseline [64.42938561167402]
We propose an online 3D reconstruction method using 3D Gaussian-based SLAM, combined with a feed-forward recurrent prediction module.<n>This approach replaces slow test-time optimization with fast network inference, significantly improving tracking speed.<n>Our method achieves performance on par with the state-of-the-art SplaTAM, while reducing tracking time by more than 90%.
arXiv Detail & Related papers (2025-08-06T16:16:58Z) - LEG-SLAM: Real-Time Language-Enhanced Gaussian Splatting for SLAM [0.0]
LEG-SLAM is a novel approach that fuses an optimized Gaussian Splatting implementation with visual-language feature extraction.<n>Our method simultaneously generates high-quality photorealistic images and semantically labeled scene maps.<n>With its potential applications in autonomous robotics, augmented reality, and other interactive domains, LEG-SLAM represents a significant step forward in real-time semantic 3D Gaussian-based SLAM.
arXiv Detail & Related papers (2025-06-03T16:51:59Z) - GSFF-SLAM: 3D Semantic Gaussian Splatting SLAM via Feature Field [17.57215792490409]
GSFF-SLAM is a novel dense semantic SLAM system based on 3D Gaussian Splatting.<n>Our method supports semantic reconstruction using various forms of 2D priors, particularly sparse and noisy signals.<n>When utilizing 2D ground truth priors, GSFF-SLAM achieves state-of-the-art semantic segmentation performance with 95.03% mIoU.
arXiv Detail & Related papers (2025-04-28T01:21:35Z) - GARAD-SLAM: 3D GAussian splatting for Real-time Anti Dynamic SLAM [9.060527946525381]
We propose GARAD-SLAM, a real-time 3DGS-based SLAM system tailored for dynamic scenes.<n>In terms of tracking, we directly perform dynamic segmentation on Gaussians and map them back to the front-end to obtain dynamic point labels.<n>Our results on real-world datasets demonstrate that our method is competitive in tracking compared to baseline methods.
arXiv Detail & Related papers (2025-02-05T14:44:17Z) - FlashSLAM: Accelerated RGB-D SLAM for Real-Time 3D Scene Reconstruction with Gaussian Splatting [14.130327598928778]
FlashSLAM is a novel SLAM approach that leverages 3D Gaussian Splatting for efficient and robust 3D scene reconstruction.<n>Existing 3DGS-based SLAM methods often fall short in sparse view settings and during large camera movements.<n>Our method achieves up to a 92% improvement in average tracking accuracy over previous methods.
arXiv Detail & Related papers (2024-12-01T05:44:38Z) - IG-SLAM: Instant Gaussian SLAM [6.228980850646457]
3D Gaussian Splatting has recently shown promising results as an alternative scene representation in SLAM systems.
We present IG-SLAM, a dense RGB-only SLAM system that employs robust Dense-SLAM methods for tracking and combines them with Gaussian Splatting.
We demonstrate competitive performance with state-of-the-art RGB-only SLAM systems while achieving faster operation speeds.
arXiv Detail & Related papers (2024-08-02T09:07:31Z) - Splat-SLAM: Globally Optimized RGB-only SLAM with 3D Gaussians [87.48403838439391]
3D Splatting has emerged as a powerful representation of geometry and appearance for RGB-only dense Simultaneous SLAM.
We propose the first RGB-only SLAM system with a dense 3D Gaussian map representation.
Our experiments on the Replica, TUM-RGBD, and ScanNet datasets indicate the effectiveness of globally optimized 3D Gaussians.
arXiv Detail & Related papers (2024-05-26T12:26:54Z) - MM3DGS SLAM: Multi-modal 3D Gaussian Splatting for SLAM Using Vision, Depth, and Inertial Measurements [59.70107451308687]
We show for the first time that using 3D Gaussians for map representation with unposed camera images and inertial measurements can enable accurate SLAM.
Our method, MM3DGS, addresses the limitations of prior rendering by enabling faster scale awareness, and improved trajectory tracking.
We also release a multi-modal dataset, UT-MM, collected from a mobile robot equipped with a camera and an inertial measurement unit.
arXiv Detail & Related papers (2024-04-01T04:57:41Z) - Loopy-SLAM: Dense Neural SLAM with Loop Closures [53.11936461015725]
We introduce Loopy-SLAM that globally optimize poses and the dense 3D model.
We use frame-to-model tracking using a data-driven point-based submap generation method and trigger loop closures online by performing global place recognition.
Evaluation on the synthetic Replica and real-world TUM-RGBD and ScanNet datasets demonstrate competitive or superior performance in tracking, mapping, and rendering accuracy when compared to existing dense neural RGBD SLAM methods.
arXiv Detail & Related papers (2024-02-14T18:18:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.