Asynchronous Multi-View SLAM
- URL: http://arxiv.org/abs/2101.06562v1
- Date: Sun, 17 Jan 2021 00:50:01 GMT
- Title: Asynchronous Multi-View SLAM
- Authors: Anqi Joyce Yang, Can Cui, Ioan Andrei B\^arsan, Raquel Urtasun,
Shenlong Wang
- Abstract summary: Existing multi-camera SLAM systems assume synchronized shutters for all cameras, which is often not the case in practice.
Our framework integrates a continuous-time motion model to relate information across asynchronous multi-frames during tracking, local mapping, and loop closing.
- Score: 78.49842639404413
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing multi-camera SLAM systems assume synchronized shutters for all
cameras, which is often not the case in practice. In this work, we propose a
generalized multi-camera SLAM formulation which accounts for asynchronous
sensor observations. Our framework integrates a continuous-time motion model to
relate information across asynchronous multi-frames during tracking, local
mapping, and loop closing. For evaluation, we collected AMV-Bench, a
challenging new SLAM dataset covering 482 km of driving recorded using our
asynchronous multi-camera robotic platform. AMV-Bench is over an order of
magnitude larger than previous multi-view HD outdoor SLAM datasets, and covers
diverse and challenging motions and environments. Our experiments emphasize the
necessity of asynchronous sensor modeling, and show that the use of multiple
cameras is critical towards robust and accurate SLAM in challenging outdoor
scenes.
Related papers
- BlinkTrack: Feature Tracking over 100 FPS via Events and Images [50.98675227695814]
We propose a novel framework, BlinkTrack, which integrates event data with RGB images for high-frequency feature tracking.
Our method extends the traditional Kalman filter into a learning-based framework, utilizing differentiable Kalman filters in both event and image branches.
Experimental results indicate that BlinkTrack significantly outperforms existing event-based methods.
arXiv Detail & Related papers (2024-09-26T15:54:18Z) - Motion Capture from Inertial and Vision Sensors [60.5190090684795]
MINIONS is a large-scale Motion capture dataset collected from INertial and visION Sensors.
We conduct experiments on multi-modal motion capture using a monocular camera and very few IMUs.
arXiv Detail & Related papers (2024-07-23T09:41:10Z) - Multicam-SLAM: Non-overlapping Multi-camera SLAM for Indirect Visual Localization and Navigation [1.3654846342364308]
This paper presents a novel approach to visual simultaneous localization and mapping (SLAM) using multiple RGB-D cameras.
The proposed method, Multicam-SLAM, significantly enhances the robustness and accuracy of SLAM systems.
Experiments in various environments demonstrate the superior accuracy and robustness of the proposed method compared to conventional single-camera SLAM systems.
arXiv Detail & Related papers (2024-06-10T15:36:23Z) - MTMMC: A Large-Scale Real-World Multi-Modal Camera Tracking Benchmark [63.878793340338035]
Multi-target multi-camera tracking is a crucial task that involves identifying and tracking individuals over time using video streams from multiple cameras.
Existing datasets for this task are either synthetically generated or artificially constructed within a controlled camera network setting.
We present MTMMC, a real-world, large-scale dataset that includes long video sequences captured by 16 multi-modal cameras in two different environments.
arXiv Detail & Related papers (2024-03-29T15:08:37Z) - VECtor: A Versatile Event-Centric Benchmark for Multi-Sensor SLAM [31.779462222706346]
Event cameras hold strong potential to complement regular cameras in situations of high dynamics or challenging illumination.
Our contribution is the first complete set of benchmark datasets captured with a multi-sensor setup.
Individual sequences include both small and large-scale environments, and cover the specific challenges targeted by dynamic vision sensors.
arXiv Detail & Related papers (2022-07-04T13:37:26Z) - Combining Events and Frames using Recurrent Asynchronous Multimodal
Networks for Monocular Depth Prediction [51.072733683919246]
We introduce Recurrent Asynchronous Multimodal (RAM) networks to handle asynchronous and irregular data from multiple sensors.
Inspired by traditional RNNs, RAM networks maintain a hidden state that is updated asynchronously and can be queried at any time to generate a prediction.
We show an improvement over state-of-the-art methods by up to 30% in terms of mean depth absolute error.
arXiv Detail & Related papers (2021-02-18T13:24:35Z) - Single-Frame based Deep View Synchronization for Unsynchronized
Multi-Camera Surveillance [56.964614522968226]
Multi-camera surveillance has been an active research topic for understanding and modeling scenes.
It is usually assumed that the cameras are all temporally synchronized when designing models for these multi-camera based tasks.
Our view synchronization models are applied to different DNNs-based multi-camera vision tasks under the unsynchronized setting.
arXiv Detail & Related papers (2020-07-08T04:39:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.