Visual Odometry for RGB-D Cameras
- URL: http://arxiv.org/abs/2203.15119v1
- Date: Mon, 28 Mar 2022 21:49:12 GMT
- Title: Visual Odometry for RGB-D Cameras
- Authors: Afonso Fontes, Jose Everardo Bessa Maia
- Abstract summary: This paper develops a quick and accurate approach to visual odometry of a moving RGB-D camera navigating on a static environment.
The proposed algorithm uses SURF as feature extractor, RANSAC to filter the results and Minimum Mean Square to estimate the rigid transformation of six parameters between successive video frames.
- Score: 3.655021726150368
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Visual odometry is the process of estimating the position and orientation of
a camera by analyzing the images associated to it. This paper develops a quick
and accurate approach to visual odometry of a moving RGB-D camera navigating on
a static environment. The proposed algorithm uses SURF (Speeded Up Robust
Features) as feature extractor, RANSAC (Random Sample Consensus) to filter the
results and Minimum Mean Square to estimate the rigid transformation of six
parameters between successive video frames. Data from a Kinect camera were used
in the tests. The results show that this approach is feasible and promising,
surpassing in performance the algorithms ICP (Interactive Closest Point) and
SfM (Structure from Motion) in tests using a publicly available dataset.
Related papers
- IMU-Aided Event-based Stereo Visual Odometry [7.280676899773076]
We improve our previous direct pipeline textitEvent-based Stereo Visual Odometry in terms of accuracy and efficiency.
To speed up the mapping operation, we propose an efficient strategy of edge-pixel sampling according to the local dynamics of events.
We release our pipeline as an open-source software for future research in this field.
arXiv Detail & Related papers (2024-05-07T07:19:25Z) - Toward Efficient Visual Gyroscopes: Spherical Moments, Harmonics Filtering, and Masking Techniques for Spherical Camera Applications [83.8743080143778]
The integration of omnidirectional cameras, offering a larger field of view compared to traditional RGB cameras, has proven to yield more accurate and robust results.
However, challenges arise in situations that lack features, have substantial noise causing significant errors, and where certain features in the images lack sufficient strength.
Here, we address these challenges by introducing a novel visual gyroscope, which combines an analytical method with a neural network approach.
arXiv Detail & Related papers (2024-04-02T13:19:06Z) - InstantSplat: Unbounded Sparse-view Pose-free Gaussian Splatting in 40 Seconds [91.77050739918037]
Novel view synthesis (NVS) from a sparse set of images has advanced significantly in 3D computer vision.
It relies on precise initial estimation of camera parameters using Structure-from-Motion (SfM)
In this study, we introduce a novel and efficient framework to enhance robust NVS from sparse-view images.
arXiv Detail & Related papers (2024-03-29T17:29:58Z) - U-ARE-ME: Uncertainty-Aware Rotation Estimation in Manhattan Environments [18.534567960292403]
We present U-ARE-ME, an algorithm that estimates camera rotation along with uncertainty from uncalibrated RGB images.
Our experiments demonstrate that U-ARE-ME performs comparably to RGB-D methods and is more robust than sparse feature-based SLAM methods.
arXiv Detail & Related papers (2024-03-22T19:14:28Z) - Anyview: Generalizable Indoor 3D Object Detection with Variable Frames [63.51422844333147]
We present a novel 3D detection framework named AnyView for our practical applications.
Our method achieves both great generalizability and high detection accuracy with a simple and clean architecture.
arXiv Detail & Related papers (2023-10-09T02:15:45Z) - RGB-based Category-level Object Pose Estimation via Decoupled Metric
Scale Recovery [72.13154206106259]
We propose a novel pipeline that decouples the 6D pose and size estimation to mitigate the influence of imperfect scales on rigid transformations.
Specifically, we leverage a pre-trained monocular estimator to extract local geometric information.
A separate branch is designed to directly recover the metric scale of the object based on category-level statistics.
arXiv Detail & Related papers (2023-09-19T02:20:26Z) - MV-ROPE: Multi-view Constraints for Robust Category-level Object Pose and Size Estimation [23.615122326731115]
We propose a novel solution that makes use of RGB video streams.
Our framework consists of three modules: a scale-aware monocular dense SLAM solution, a lightweight object pose predictor, and an object-level pose graph.
Our experimental results demonstrate that when utilizing public dataset sequences with high-quality depth information, the proposed method exhibits comparable performance to state-of-the-art RGB-D methods.
arXiv Detail & Related papers (2023-08-17T08:29:54Z) - A Combined Approach Toward Consistent Reconstructions of Indoor Spaces
Based on 6D RGB-D Odometry and KinectFusion [7.503338065129185]
We propose a 6D RGB-D odometry approach that finds the relative camera pose between consecutive RGB-D frames by keypoint extraction.
We feed the estimated pose to the highly accurate KinectFusion algorithm, which fine-tune the frame-to-frame relative pose.
Our algorithm outputs a ready-to-use polygon mesh (highly suitable for creating 3D virtual worlds) without any postprocessing steps.
arXiv Detail & Related papers (2022-12-25T22:52:25Z) - ParticleSfM: Exploiting Dense Point Trajectories for Localizing Moving
Cameras in the Wild [57.37891682117178]
We present a robust dense indirect structure-from-motion method for videos that is based on dense correspondence from pairwise optical flow.
A novel neural network architecture is proposed for processing irregular point trajectory data.
Experiments on MPI Sintel dataset show that our system produces significantly more accurate camera trajectories.
arXiv Detail & Related papers (2022-07-19T09:19:45Z) - Moving object detection for visual odometry in a dynamic environment
based on occlusion accumulation [31.143322364794894]
We propose a moving object detection algorithm that uses RGB-D images.
The proposed algorithm does not require estimating a background model.
We use dense visual odometry (DVO) as a VO method with a bi-square regression weight.
arXiv Detail & Related papers (2020-09-18T11:01:46Z) - Multi-View Photometric Stereo: A Robust Solution and Benchmark Dataset
for Spatially Varying Isotropic Materials [65.95928593628128]
We present a method to capture both 3D shape and spatially varying reflectance with a multi-view photometric stereo technique.
Our algorithm is suitable for perspective cameras and nearby point light sources.
arXiv Detail & Related papers (2020-01-18T12:26:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.