Comparative Study of Visual SLAM-Based Mobile Robot Localization Using
Fiducial Markers
- URL: http://arxiv.org/abs/2309.04441v1
- Date: Fri, 8 Sep 2023 17:05:24 GMT
- Title: Comparative Study of Visual SLAM-Based Mobile Robot Localization Using
Fiducial Markers
- Authors: Jongwon Lee, Su Yeon Choi, David Hanley, Timothy Bretl
- Abstract summary: This paper presents a comparative study of three modes for mobile robot localization based on visual SLAM using fiducial markers.
The reason for comparing the SLAM-based approaches is because previous work has shown their superior performance over feature-only methods.
Hardware experiments show consistent trajectory error levels across the three modes, with the localization mode exhibiting the shortest runtime among them.
- Score: 4.918853205874711
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper presents a comparative study of three modes for mobile robot
localization based on visual SLAM using fiducial markers (i.e., square-shaped
artificial landmarks with a black-and-white grid pattern): SLAM, SLAM with a
prior map, and localization with a prior map. The reason for comparing the
SLAM-based approaches leveraging fiducial markers is because previous work has
shown their superior performance over feature-only methods, with less
computational burden compared to methods that use both feature and marker
detection without compromising the localization performance. The evaluation is
conducted using indoor image sequences captured with a hand-held camera
containing multiple fiducial markers in the environment. The performance
metrics include absolute trajectory error and runtime for the optimization
process per frame. In particular, for the last two modes (SLAM and localization
with a prior map), we evaluate their performances by perturbing the quality of
prior map to study the extent to which each mode is tolerant to such
perturbations. Hardware experiments show consistent trajectory error levels
across the three modes, with the localization mode exhibiting the shortest
runtime among them. Yet, with map perturbations, SLAM with a prior map
maintains performance, while localization mode degrades in both aspects.
Related papers
- Loopy-SLAM: Dense Neural SLAM with Loop Closures [53.11936461015725]
We introduce Loopy-SLAM that globally optimize poses and the dense 3D model.
We use frame-to-model tracking using a data-driven point-based submap generation method and trigger loop closures online by performing global place recognition.
Evaluation on the synthetic Replica and real-world TUM-RGBD and ScanNet datasets demonstrate competitive or superior performance in tracking, mapping, and rendering accuracy when compared to existing dense neural RGBD SLAM methods.
arXiv Detail & Related papers (2024-02-14T18:18:32Z) - Learning Saliency From Fixations [0.9208007322096533]
We present a novel approach for saliency prediction in images, leveraging parallel decoding in transformers to learn saliency solely from fixation maps.
Our approach, named Saliency TRansformer (SalTR), achieves metric scores on par with state-of-the-art approaches on the Salicon and MIT300 benchmarks.
arXiv Detail & Related papers (2023-11-23T16:04:41Z) - GS-SLAM: Dense Visual SLAM with 3D Gaussian Splatting [51.96353586773191]
We introduce textbfGS-SLAM that first utilizes 3D Gaussian representation in the Simultaneous Localization and Mapping system.
Our method utilizes a real-time differentiable splatting rendering pipeline that offers significant speedup to map optimization and RGB-D rendering.
Our method achieves competitive performance compared with existing state-of-the-art real-time methods on the Replica, TUM-RGBD datasets.
arXiv Detail & Related papers (2023-11-20T12:08:23Z) - Volumetric Semantically Consistent 3D Panoptic Mapping [77.13446499924977]
We introduce an online 2D-to-3D semantic instance mapping algorithm aimed at generating semantic 3D maps suitable for autonomous agents in unstructured environments.
It introduces novel ways of integrating semantic prediction confidence during mapping, producing semantic and instance-consistent 3D regions.
The proposed method achieves accuracy superior to the state of the art on public large-scale datasets, improving on a number of widely used metrics.
arXiv Detail & Related papers (2023-09-26T08:03:10Z) - A Locality-based Neural Solver for Optical Motion Capture [37.28597049192196]
Given noisy marker data, we propose a new heterogeneous graph neural network which treats markers and joints as different types of nodes.
We show that our method outperforms state-of-the-art methods in terms of prediction accuracy of occluded marker position error.
arXiv Detail & Related papers (2023-09-01T12:40:17Z) - Boosting Few-shot Fine-grained Recognition with Background Suppression
and Foreground Alignment [53.401889855278704]
Few-shot fine-grained recognition (FS-FGR) aims to recognize novel fine-grained categories with the help of limited available samples.
We propose a two-stage background suppression and foreground alignment framework, which is composed of a background activation suppression (BAS) module, a foreground object alignment (FOA) module, and a local to local (L2L) similarity metric.
Experiments conducted on multiple popular fine-grained benchmarks demonstrate that our method outperforms the existing state-of-the-art by a large margin.
arXiv Detail & Related papers (2022-10-04T07:54:40Z) - FD-SLAM: 3-D Reconstruction Using Features and Dense Matching [18.577229381683434]
We propose an RGB-D SLAM system that uses dense frame-to-model odometry to build accurate sub-maps.
We incorporate a learning-based loop closure component based on 3-D features which further stabilises map building.
The approach can also scale to large scenes where other systems often fail.
arXiv Detail & Related papers (2022-03-25T18:58:46Z) - CAMERAS: Enhanced Resolution And Sanity preserving Class Activation
Mapping for image saliency [61.40511574314069]
Backpropagation image saliency aims at explaining model predictions by estimating model-centric importance of individual pixels in the input.
We propose CAMERAS, a technique to compute high-fidelity backpropagation saliency maps without requiring any external priors.
arXiv Detail & Related papers (2021-06-20T08:20:56Z) - Improved Real-Time Monocular SLAM Using Semantic Segmentation on
Selective Frames [15.455647477995312]
monocular simultaneous localization and mapping (SLAM) is emerging in advanced driver assistance systems and autonomous driving.
This paper proposes an improved real-time monocular SLAM using deep learning-based semantic segmentation.
Experiments with six video sequences demonstrate that the proposed monocular SLAM system achieves significantly more accurate trajectory tracking accuracy.
arXiv Detail & Related papers (2021-04-30T22:34:45Z) - Accurate Visual-Inertial SLAM by Feature Re-identification [4.263022790692934]
We propose an efficient drift-less SLAM method by re-identifying existing features from a spatial-temporal sensitive sub-global map.
Our method achieves 67.3% and 87.5% absolute translation error reduction with only a small additional computational cost.
arXiv Detail & Related papers (2021-02-26T12:54:33Z) - Making Affine Correspondences Work in Camera Geometry Computation [62.7633180470428]
Local features provide region-to-region rather than point-to-point correspondences.
We propose guidelines for effective use of region-to-region matches in the course of a full model estimation pipeline.
Experiments show that affine solvers can achieve accuracy comparable to point-based solvers at faster run-times.
arXiv Detail & Related papers (2020-07-20T12:07:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.