DynaSLAM II: Tightly-Coupled Multi-Object Tracking and SLAM
- URL: http://arxiv.org/abs/2010.07820v1
- Date: Thu, 15 Oct 2020 15:25:30 GMT
- Title: DynaSLAM II: Tightly-Coupled Multi-Object Tracking and SLAM
- Authors: Berta Bescos, Carlos Campos, Juan D. Tard\'os, Jos\'e Neira
- Abstract summary: DynaSLAM II is a visual SLAM system for stereo and RGB-D configurations that tightly integrates the multi-object tracking capability.
We demonstrate that tracking dynamic objects does not only provide rich clues for scene understanding but is also beneficial for camera tracking.
- Score: 2.9822184411723645
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The assumption of scene rigidity is common in visual SLAM algorithms.
However, it limits their applicability in populated real-world environments.
Furthermore, most scenarios including autonomous driving, multi-robot
collaboration and augmented/virtual reality, require explicit motion
information of the surroundings to help with decision making and scene
understanding. We present in this paper DynaSLAM II, a visual SLAM system for
stereo and RGB-D configurations that tightly integrates the multi-object
tracking capability.
DynaSLAM II makes use of instance semantic segmentation and of ORB features
to track dynamic objects. The structure of the static scene and of the dynamic
objects is optimized jointly with the trajectories of both the camera and the
moving agents within a novel bundle adjustment proposal. The 3D bounding boxes
of the objects are also estimated and loosely optimized within a fixed temporal
window. We demonstrate that tracking dynamic objects does not only provide rich
clues for scene understanding but is also beneficial for camera tracking.
The project code will be released upon acceptance.
Related papers
- Shape of Motion: 4D Reconstruction from a Single Video [51.04575075620677]
We introduce a method capable of reconstructing generic dynamic scenes, featuring explicit, full-sequence-long 3D motion.
We exploit the low-dimensional structure of 3D motion by representing scene motion with a compact set of SE3 motion bases.
Our method achieves state-of-the-art performance for both long-range 3D/2D motion estimation and novel view synthesis on dynamic scenes.
arXiv Detail & Related papers (2024-07-18T17:59:08Z) - HUGS: Holistic Urban 3D Scene Understanding via Gaussian Splatting [53.6394928681237]
holistic understanding of urban scenes based on RGB images is a challenging yet important problem.
Our main idea involves the joint optimization of geometry, appearance, semantics, and motion using a combination of static and dynamic 3D Gaussians.
Our approach offers the ability to render new viewpoints in real-time, yielding 2D and 3D semantic information with high accuracy.
arXiv Detail & Related papers (2024-03-19T13:39:05Z) - DrivingGaussian: Composite Gaussian Splatting for Surrounding Dynamic Autonomous Driving Scenes [57.12439406121721]
We present DrivingGaussian, an efficient and effective framework for surrounding dynamic autonomous driving scenes.
For complex scenes with moving objects, we first sequentially and progressively model the static background of the entire scene.
We then leverage a composite dynamic Gaussian graph to handle multiple moving objects.
We further use a LiDAR prior for Gaussian Splatting to reconstruct scenes with greater details and maintain panoramic consistency.
arXiv Detail & Related papers (2023-12-13T06:30:51Z) - MOTSLAM: MOT-assisted monocular dynamic SLAM using single-view depth
estimation [5.33931801679129]
MOTSLAM is a dynamic visual SLAM system with the monocular configuration that tracks both poses and bounding boxes of dynamic objects.
Our experiments on the KITTI dataset demonstrate that our system has reached best performance on both camera ego-motion and object tracking on monocular dynamic SLAM.
arXiv Detail & Related papers (2022-10-05T06:07:10Z) - TwistSLAM++: Fusing multiple modalities for accurate dynamic semantic
SLAM [0.0]
TwistSLAM++ is a semantic, dynamic, SLAM system that fuses stereo images and LiDAR information.
We show on classical benchmarks that this fusion approach based on multimodal information improves the accuracy of object tracking.
arXiv Detail & Related papers (2022-09-16T12:28:21Z) - TwistSLAM: Constrained SLAM in Dynamic Environment [0.0]
We present TwistSLAM, a semantic, dynamic, stereo SLAM system that can track dynamic objects in the scene.
Our algorithm creates clusters of points according to their semantic class.
It uses the static parts of the environment to robustly localize the camera and tracks the remaining objects.
arXiv Detail & Related papers (2022-02-24T22:08:45Z) - DL-SLOT: Dynamic Lidar SLAM and Object Tracking Based On Graph
Optimization [2.889268075288957]
Ego-pose estimation and dynamic object tracking are two key issues in an autonomous driving system.
In this paper, DL-SLOT, a dynamic Lidar SLAM and object tracking method is proposed.
We perform SLAM and object tracking simultaneously in this framework, which significantly improves the robustness and accuracy of SLAM in highly dynamic road scenarios.
arXiv Detail & Related papers (2022-02-23T11:22:43Z) - Class-agnostic Reconstruction of Dynamic Objects from Videos [127.41336060616214]
We introduce REDO, a class-agnostic framework to REconstruct the Dynamic Objects from RGBD or calibrated videos.
We develop two novel modules. First, we introduce a canonical 4D implicit function which is pixel-aligned with aggregated temporal visual cues.
Second, we develop a 4D transformation module which captures object dynamics to support temporal propagation and aggregation.
arXiv Detail & Related papers (2021-12-03T18:57:47Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z) - DOT: Dynamic Object Tracking for Visual SLAM [83.69544718120167]
DOT combines instance segmentation and multi-view geometry to generate masks for dynamic objects.
To determine which objects are actually moving, DOT segments first instances of potentially dynamic objects and then, with the estimated camera motion, tracks such objects by minimizing the photometric reprojection error.
Our results show that our approach improves significantly the accuracy and robustness of ORB-SLAM 2, especially in highly dynamic scenes.
arXiv Detail & Related papers (2020-09-30T18:36:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.