IBISCape: A Simulated Benchmark for multi-modal SLAM Systems Evaluation
in Large-scale Dynamic Environments
- URL: http://arxiv.org/abs/2206.13455v2
- Date: Thu, 20 Oct 2022 09:06:07 GMT
- Title: IBISCape: A Simulated Benchmark for multi-modal SLAM Systems Evaluation
in Large-scale Dynamic Environments
- Authors: Abanob Soliman, Fabien Bonardi, D\'esir\'e Sidib\'e and Samia Bouchafa
- Abstract summary: IBISCape is a simulated benchmark for high-fidelity SLAM systems.
We offer 34 multi-modal datasets suitable for autonomous vehicles navigation.
We evaluate four ORB-SLAM3 systems on various sequences collected in simulated large-scale dynamic environments.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The development process of high-fidelity SLAM systems depends on their
validation upon reliable datasets. Towards this goal, we propose IBISCape, a
simulated benchmark that includes data synchronization and acquisition APIs for
telemetry from heterogeneous sensors: stereo-RGB/DVS, Depth, IMU, and GPS,
along with the ground truth scene segmentation and vehicle ego-motion. Our
benchmark is built upon the CARLA simulator, whose back-end is the Unreal
Engine rendering a high dynamic scenery simulating the real world. Moreover, we
offer 34 multi-modal datasets suitable for autonomous vehicles navigation,
including scenarios for scene understanding evaluation like accidents, along
with a wide range of frame quality based on a dynamic weather simulation class
integrated with our APIs. We also introduce the first calibration targets to
CARLA maps to solve the unknown distortion parameters problem of CARLA
simulated DVS and RGB cameras. Finally, using IBISCape sequences, we evaluate
four ORB-SLAM3 systems (monocular RGB, stereo RGB, Stereo Visual Inertial
(SVI), and RGB-D) performance and BASALT Visual-Inertial Odometry (VIO) system
on various sequences collected in simulated large-scale dynamic environments.
Keywords: benchmark, multi-modal, datasets, Odometry, Calibration, DVS, SLAM
Related papers
- DrivingSphere: Building a High-fidelity 4D World for Closed-loop Simulation [54.02069690134526]
We propose DrivingSphere, a realistic and closed-loop simulation framework.
Its core idea is to build 4D world representation and generate real-life and controllable driving scenarios.
By providing a dynamic and realistic simulation environment, DrivingSphere enables comprehensive testing and validation of autonomous driving algorithms.
arXiv Detail & Related papers (2024-11-18T03:00:33Z) - ES-PTAM: Event-based Stereo Parallel Tracking and Mapping [11.801511288805225]
Event cameras offer advantages to overcome the limitations of standard cameras.
We propose a novel event-based stereo VO system by combining two ideas.
We evaluate the system on five real-world datasets.
arXiv Detail & Related papers (2024-08-28T07:56:28Z) - From Perfect to Noisy World Simulation: Customizable Embodied Multi-modal Perturbations for SLAM Robustness Benchmarking [32.52171076424419]
Embodied agents require robust navigation systems to operate in unstructured environments.
We propose a novel, customizable pipeline for noisy data synthesis.
Our analysis uncovers the susceptibilities of both neural (NeRF) and non-neural SLAM models to disturbances.
arXiv Detail & Related papers (2024-06-24T17:57:05Z) - SEVD: Synthetic Event-based Vision Dataset for Ego and Fixed Traffic Perception [22.114089372056238]
We present SEVD, a first-of-its-kind multi-view ego, and fixed perception synthetic event-based dataset.
SEVD spans urban, suburban, rural, and highway scenes featuring various classes of objects.
We evaluate the dataset using state-of-the-art event-based (RED, RVT) and frame-based (YOLOv8) methods for traffic participant detection tasks.
arXiv Detail & Related papers (2024-04-12T20:40:12Z) - Deep Generative Data Assimilation in Multimodal Setting [0.1052166918701117]
In this work, we propose SLAMS: Score-based Latent Assimilation in Multimodal Setting.
We assimilate in-situ weather station data and ex-situ satellite imagery to calibrate the vertical temperature profiles, globally.
Our work is the first to apply deep generative framework for multimodal data assimilation using real-world datasets.
arXiv Detail & Related papers (2024-04-10T00:25:09Z) - MM3DGS SLAM: Multi-modal 3D Gaussian Splatting for SLAM Using Vision, Depth, and Inertial Measurements [59.70107451308687]
We show for the first time that using 3D Gaussians for map representation with unposed camera images and inertial measurements can enable accurate SLAM.
Our method, MM3DGS, addresses the limitations of prior rendering by enabling faster scale awareness, and improved trajectory tracking.
We also release a multi-modal dataset, UT-MM, collected from a mobile robot equipped with a camera and an inertial measurement unit.
arXiv Detail & Related papers (2024-04-01T04:57:41Z) - DH-PTAM: A Deep Hybrid Stereo Events-Frames Parallel Tracking And Mapping System [1.443696537295348]
This paper presents a robust approach for a visual parallel tracking and mapping (PTAM) system that excels in challenging environments.
Our proposed method combines the strengths of heterogeneous multi-modal visual sensors, in a unified reference frame.
Our implementation's research-based Python API is publicly available on GitHub.
arXiv Detail & Related papers (2023-06-02T19:52:13Z) - 4Seasons: Benchmarking Visual SLAM and Long-Term Localization for
Autonomous Driving in Challenging Conditions [54.59279160621111]
We present a novel visual SLAM and long-term localization benchmark for autonomous driving in challenging conditions based on the large-scale 4Seasons dataset.
The proposed benchmark provides drastic appearance variations caused by seasonal changes and diverse weather and illumination conditions.
We introduce a new unified benchmark for jointly evaluating visual odometry, global place recognition, and map-based visual localization performance.
arXiv Detail & Related papers (2022-12-31T13:52:36Z) - Using Detection, Tracking and Prediction in Visual SLAM to Achieve
Real-time Semantic Mapping of Dynamic Scenarios [70.70421502784598]
RDS-SLAM can build semantic maps at object level for dynamic scenarios in real time using only one commonly used Intel Core i7 CPU.
We evaluate RDS-SLAM in TUM RGB-D dataset, and experimental results show that RDS-SLAM can run with 30.3 ms per frame in dynamic scenarios.
arXiv Detail & Related papers (2022-10-10T11:03:32Z) - Learning to Simulate Realistic LiDARs [66.7519667383175]
We introduce a pipeline for data-driven simulation of a realistic LiDAR sensor.
We show that our model can learn to encode realistic effects such as dropped points on transparent surfaces.
We use our technique to learn models of two distinct LiDAR sensors and use them to improve simulated LiDAR data accordingly.
arXiv Detail & Related papers (2022-09-22T13:12:54Z) - Towards Scale Consistent Monocular Visual Odometry by Learning from the
Virtual World [83.36195426897768]
We propose VRVO, a novel framework for retrieving the absolute scale from virtual data.
We first train a scale-aware disparity network using both monocular real images and stereo virtual data.
The resulting scale-consistent disparities are then integrated with a direct VO system.
arXiv Detail & Related papers (2022-03-11T01:51:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.