Dynamic Dense RGB-D SLAM using Learning-based Visual Odometry
- URL: http://arxiv.org/abs/2205.05916v2
- Date: Wed, 28 Jun 2023 20:31:20 GMT
- Title: Dynamic Dense RGB-D SLAM using Learning-based Visual Odometry
- Authors: Shihao Shen, Yilin Cai, Jiayi Qiu, Guangzhao Li
- Abstract summary: We propose a dense dynamic RGB-D SLAM pipeline based on a learning-based visual odometry, TartanVO.
Our pipeline resolves dynamic/static segmentation by leveraging the optical flow output, and only fuse static points into the map.
- Score: 0.8029049649310211
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We propose a dense dynamic RGB-D SLAM pipeline based on a learning-based
visual odometry, TartanVO. TartanVO, like other direct methods rather than
feature-based, estimates camera pose through dense optical flow, which only
applies to static scenes and disregards dynamic objects. Due to the color
constancy assumption, optical flow is not able to differentiate between dynamic
and static pixels. Therefore, to reconstruct a static map through such direct
methods, our pipeline resolves dynamic/static segmentation by leveraging the
optical flow output, and only fuse static points into the map. Moreover, we
rerender the input frames such that the dynamic pixels are removed and
iteratively pass them back into the visual odometry to refine the pose
estimate.
Related papers
- V3D-SLAM: Robust RGB-D SLAM in Dynamic Environments with 3D Semantic Geometry Voting [1.3493547928462395]
Simultaneous localization and mapping (SLAM) in highly dynamic environments is challenging due to the correlation between moving objects and the camera pose.
We propose a robust method, V3D-SLAM, to remove moving objects via two lightweight re-evaluation stages.
Our experiment on the TUM RGB-D benchmark on dynamic sequences with ground-truth camera trajectories showed that our methods outperform the most recent state-of-the-art SLAM methods.
arXiv Detail & Related papers (2024-10-15T21:08:08Z) - Dynamic in Static: Hybrid Visual Correspondence for Self-Supervised Video Object Segmentation [126.12940972028012]
We present HVC, a framework for self-supervised video object segmentation.
HVC extracts pseudo-dynamic signals from static images, enabling an efficient and scalable VOS model.
We propose a hybrid visual correspondence loss to learn joint static and dynamic consistency representations.
arXiv Detail & Related papers (2024-04-21T02:21:30Z) - Vision-Informed Flow Image Super-Resolution with Quaternion Spatial
Modeling and Dynamic Flow Convolution [49.45309818782329]
Flow image super-resolution (FISR) aims at recovering high-resolution turbulent velocity fields from low-resolution flow images.
Existing FISR methods mainly process the flow images in natural image patterns.
We propose the first flow visual property-informed FISR algorithm.
arXiv Detail & Related papers (2024-01-29T06:48:16Z) - MPI-Flow: Learning Realistic Optical Flow with Multiplane Images [18.310665144874775]
We investigate generating realistic optical flow datasets from real-world images.
To generate highly realistic new images, we construct a layered depth representation, known as multiplane images (MPI), from single-view images.
To ensure the realism of motion, we present an independent object motion module that can separate the camera and dynamic object motion in MPI.
arXiv Detail & Related papers (2023-09-13T04:31:00Z) - Adaptive Multi-source Predictor for Zero-shot Video Object Segmentation [68.56443382421878]
We propose a novel adaptive multi-source predictor for zero-shot video object segmentation (ZVOS)
In the static object predictor, the RGB source is converted to depth and static saliency sources, simultaneously.
Experiments show that the proposed model outperforms the state-of-the-art methods on three challenging ZVOS benchmarks.
arXiv Detail & Related papers (2023-03-18T10:19:29Z) - Real-time SLAM Pipeline in Dynamics Environment [0.0]
We are presenting a solution which use RGB-D SLAM as well as YOLO real-time object detection to segment and remove dynamic scene and then construct static scene 3D.
We gathered a dataset which allows us to jointly consider semantics, geometry, and physics and thus enables us to reconstruct the static scene while filtering out all dynamic objects.
arXiv Detail & Related papers (2023-03-04T00:08:52Z) - Deep Dynamic Scene Deblurring from Optical Flow [53.625999196063574]
Deblurring can provide visually more pleasant pictures and make photography more convenient.
It is difficult to model the non-uniform blur mathematically.
We develop a convolutional neural network (CNN) to restore the sharp images from the deblurred features.
arXiv Detail & Related papers (2023-01-18T06:37:21Z) - Physics Informed Neural Fields for Smoke Reconstruction with Sparse Data [73.8970871148949]
High-fidelity reconstruction of fluids from sparse multiview RGB videos remains a formidable challenge.
Existing solutions either assume knowledge of obstacles and lighting, or only focus on simple fluid scenes without obstacles or complex lighting.
We present the first method to reconstruct dynamic fluid by leveraging the governing physics (ie, Navier -Stokes equations) in an end-to-end optimization.
arXiv Detail & Related papers (2022-06-14T03:38:08Z) - DOT: Dynamic Object Tracking for Visual SLAM [83.69544718120167]
DOT combines instance segmentation and multi-view geometry to generate masks for dynamic objects.
To determine which objects are actually moving, DOT segments first instances of potentially dynamic objects and then, with the estimated camera motion, tracks such objects by minimizing the photometric reprojection error.
Our results show that our approach improves significantly the accuracy and robustness of ORB-SLAM 2, especially in highly dynamic scenes.
arXiv Detail & Related papers (2020-09-30T18:36:28Z) - Dynamic Object Tracking and Masking for Visual SLAM [1.37013665345905]
In dynamic environments, performance of visual SLAM techniques can be impaired by visual features taken from moving objects.
This paper presents a pipeline that uses deep neural networks, extended Kalman filters and visual SLAM to improve both localization and mapping.
arXiv Detail & Related papers (2020-07-31T20:37:14Z) - FlowFusion: Dynamic Dense RGB-D SLAM Based on Optical Flow [17.040818114071833]
We present a novel dense RGB-D SLAM solution that simultaneously accomplishes the dynamic/static segmentation and camera ego-motion estimation.
Our novelty is using optical flow residuals to highlight the dynamic semantics in the RGB-D point clouds.
arXiv Detail & Related papers (2020-03-11T04:00:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.