RP-VIO: Robust Plane-based Visual-Inertial Odometry for Dynamic
Environments
- URL: http://arxiv.org/abs/2103.10400v1
- Date: Thu, 18 Mar 2021 17:33:07 GMT
- Title: RP-VIO: Robust Plane-based Visual-Inertial Odometry for Dynamic
Environments
- Authors: Karnik Ram, Chaitanya Kharyal, Sudarshan S. Harithas, K. Madhava
Krishna
- Abstract summary: We present RP-VIO, a state-of-the-art visual-inertial odometry system for dynamic environments.
We also present a highly-dynamic, photorealistic synthetic dataset for a more effective evaluation of the capabilities of modern VINS systems.
- Score: 14.260575326111585
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern visual-inertial navigation systems (VINS) are faced with a critical
challenge in real-world deployment: they need to operate reliably and robustly
in highly dynamic environments. Current best solutions merely filter dynamic
objects as outliers based on the semantics of the object category. Such an
approach does not scale as it requires semantic classifiers to encompass all
possibly-moving object classes; this is hard to define, let alone deploy. On
the other hand, many real-world environments exhibit strong structural
regularities in the form of planes such as walls and ground surfaces, which are
also crucially static. We present RP-VIO, a monocular visual-inertial odometry
system that leverages the simple geometry of these planes for improved
robustness and accuracy in challenging dynamic environments. Since existing
datasets have a limited number of dynamic elements, we also present a
highly-dynamic, photorealistic synthetic dataset for a more effective
evaluation of the capabilities of modern VINS systems. We evaluate our approach
on this dataset, and three diverse sequences from standard datasets including
two real-world dynamic sequences and show a significant improvement in
robustness and accuracy over a state-of-the-art monocular visual-inertial
odometry system. We also show in simulation an improvement over a simple
dynamic-features masking approach. Our code and dataset are publicly available.
Related papers
- GMS-VINS:Multi-category Dynamic Objects Semantic Segmentation for Enhanced Visual-Inertial Odometry Using a Promptable Foundation Model [7.07379964916809]
We introduce GMS-VINS, which integrates an enhanced SORT algorithm along with a robust multi-category segmentation framework into visual-inertial odometry (VIO)
The enhanced SORT algorithm significantly improves the reliability of tracking multiple dynamic objects.
Our proposed method performs impressively in multiple scenarios, outperforming other state-of-the-art methods.
arXiv Detail & Related papers (2024-11-28T17:41:33Z) - OmniPose6D: Towards Short-Term Object Pose Tracking in Dynamic Scenes from Monocular RGB [40.62577054196799]
We introduce a large-scale synthetic dataset OmniPose6D, crafted to mirror the diversity of real-world conditions.
We present a benchmarking framework for a comprehensive comparison of pose tracking algorithms.
arXiv Detail & Related papers (2024-10-09T09:01:40Z) - MonST3R: A Simple Approach for Estimating Geometry in the Presence of Motion [118.74385965694694]
We present Motion DUSt3R (MonST3R), a novel geometry-first approach that directly estimates per-timestep geometry from dynamic scenes.
By simply estimating a pointmap for each timestep, we can effectively adapt DUST3R's representation, previously only used for static scenes, to dynamic scenes.
We show that by posing the problem as a fine-tuning task, identifying several suitable datasets, and strategically training the model on this limited data, we can surprisingly enable the model to handle dynamics.
arXiv Detail & Related papers (2024-10-04T18:00:07Z) - DENSER: 3D Gaussians Splatting for Scene Reconstruction of Dynamic Urban Environments [0.0]
We propose DENSER, a framework that significantly enhances the representation of dynamic objects.
The proposed approach significantly outperforms state-of-the-art methods by a wide margin.
arXiv Detail & Related papers (2024-09-16T07:11:58Z) - Dynamic Scene Understanding through Object-Centric Voxelization and Neural Rendering [57.895846642868904]
We present a 3D generative model named DynaVol-S for dynamic scenes that enables object-centric learning.
voxelization infers per-object occupancy probabilities at individual spatial locations.
Our approach integrates 2D semantic features to create 3D semantic grids, representing the scene through multiple disentangled voxel grids.
arXiv Detail & Related papers (2024-07-30T15:33:58Z) - Visual-Inertial Multi-Instance Dynamic SLAM with Object-level
Relocalisation [14.302118093865849]
We present a tightly-coupled visual-inertial object-level multi-instance dynamic SLAM system.
It can robustly optimise for the camera pose, velocity, IMU biases and build a dense 3D reconstruction object-level map of the environment.
arXiv Detail & Related papers (2022-08-08T17:13:24Z) - ACID: Action-Conditional Implicit Visual Dynamics for Deformable Object
Manipulation [135.10594078615952]
We introduce ACID, an action-conditional visual dynamics model for volumetric deformable objects.
A benchmark contains over 17,000 action trajectories with six types of plush toys and 78 variants.
Our model achieves the best performance in geometry, correspondence, and dynamics predictions.
arXiv Detail & Related papers (2022-03-14T04:56:55Z) - Towards Scale Consistent Monocular Visual Odometry by Learning from the
Virtual World [83.36195426897768]
We propose VRVO, a novel framework for retrieving the absolute scale from virtual data.
We first train a scale-aware disparity network using both monocular real images and stereo virtual data.
The resulting scale-consistent disparities are then integrated with a direct VO system.
arXiv Detail & Related papers (2022-03-11T01:51:54Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - DOT: Dynamic Object Tracking for Visual SLAM [83.69544718120167]
DOT combines instance segmentation and multi-view geometry to generate masks for dynamic objects.
To determine which objects are actually moving, DOT segments first instances of potentially dynamic objects and then, with the estimated camera motion, tracks such objects by minimizing the photometric reprojection error.
Our results show that our approach improves significantly the accuracy and robustness of ORB-SLAM 2, especially in highly dynamic scenes.
arXiv Detail & Related papers (2020-09-30T18:36:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.