Stereo Visual Odometry with Deep Learning-Based Point and Line Feature
Matching using an Attention Graph Neural Network
- URL: http://arxiv.org/abs/2308.01125v1
- Date: Wed, 2 Aug 2023 13:09:12 GMT
- Title: Stereo Visual Odometry with Deep Learning-Based Point and Line Feature
Matching using an Attention Graph Neural Network
- Authors: Shenbagaraj Kannapiran, Nalin Bendapudi, Ming-Yuan Yu, Devarth Parikh,
Spring Berman, Ankit Vora, and Gaurav Pandey
- Abstract summary: We present a Stereo Visual Odometry (StereoVO) technique based on point and line features.
Our method achieves more line feature matches than state-of-the-art line matching algorithms.
- Score: 4.900326352413157
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Robust feature matching forms the backbone for most Visual Simultaneous
Localization and Mapping (vSLAM), visual odometry, 3D reconstruction, and
Structure from Motion (SfM) algorithms. However, recovering feature matches
from texture-poor scenes is a major challenge and still remains an open area of
research. In this paper, we present a Stereo Visual Odometry (StereoVO)
technique based on point and line features which uses a novel feature-matching
mechanism based on an Attention Graph Neural Network that is designed to
perform well even under adverse weather conditions such as fog, haze, rain, and
snow, and dynamic lighting conditions such as nighttime illumination and glare
scenarios. We perform experiments on multiple real and synthetic datasets to
validate the ability of our method to perform StereoVO under low visibility
weather and lighting conditions through robust point and line matches. The
results demonstrate that our method achieves more line feature matches than
state-of-the-art line matching algorithms, which when complemented with point
feature matches perform consistently well in adverse weather and dynamic
lighting conditions.
Related papers
- DeepArUco++: Improved detection of square fiducial markers in challenging lighting conditions [3.783609886054562]
Fiducial markers are a computer vision tool used for object pose estimation and detection.
DeepArUco++ is a framework that performs marker detection and decoding in challenging lighting conditions.
We present a second, real-life dataset of ArUco markers in challenging lighting conditions used to evaluate our system.
arXiv Detail & Related papers (2024-11-08T13:18:31Z) - A Radiometric Correction based Optical Modeling Approach to Removing Reflection Noise in TLS Point Clouds of Urban Scenes [3.7967365472200894]
TLS-acquired point clouds often contain virtual points from reflective surfaces, causing disruptions.
This study presents a reflection noise elimination algorithm for TLS point clouds.
arXiv Detail & Related papers (2024-07-03T06:17:41Z) - NeRF-VO: Real-Time Sparse Visual Odometry with Neural Radiance Fields [13.178099653374945]
NeRF-VO integrates learning-based sparse visual odometry for low-latency camera tracking and a neural radiance scene representation.
We surpass SOTA methods in pose estimation accuracy, novel view fidelity, and dense reconstruction quality across a variety of synthetic and real-world datasets.
arXiv Detail & Related papers (2023-12-20T22:42:17Z) - DNS SLAM: Dense Neural Semantic-Informed SLAM [92.39687553022605]
DNS SLAM is a novel neural RGB-D semantic SLAM approach featuring a hybrid representation.
Our method integrates multi-view geometry constraints with image-based feature extraction to improve appearance details.
Our experimental results achieve state-of-the-art performance on both synthetic data and real-world data tracking.
arXiv Detail & Related papers (2023-11-30T21:34:44Z) - Spatiotemporally Consistent HDR Indoor Lighting Estimation [66.26786775252592]
We propose a physically-motivated deep learning framework to solve the indoor lighting estimation problem.
Given a single LDR image with a depth map, our method predicts spatially consistent lighting at any given image position.
Our framework achieves photorealistic lighting prediction with higher quality compared to state-of-the-art single-image or video-based methods.
arXiv Detail & Related papers (2023-05-07T20:36:29Z) - Local Feature Extraction from Salient Regions by Feature Map
Transformation [0.7734726150561086]
We propose a framework that robustly extracts and describes salient local features regardless of changing light and viewpoints.
The framework suppresses illumination variations and encourages structural information to ignore the noise from light.
Our model extracts feature points from salient regions leading to reduced incorrect matches.
arXiv Detail & Related papers (2023-01-25T05:31:20Z) - Real-Time Simultaneous Localization and Mapping with LiDAR intensity [9.374695605941627]
We propose a novel real-time LiDAR intensity image-based simultaneous localization and mapping method.
Our method can run in real time with high accuracy and works well with illumination changes, low-texture, and unstructured environments.
arXiv Detail & Related papers (2023-01-23T03:59:48Z) - Progressively-connected Light Field Network for Efficient View Synthesis [69.29043048775802]
We present a Progressively-connected Light Field network (ProLiF) for the novel view synthesis of complex forward-facing scenes.
ProLiF encodes a 4D light field, which allows rendering a large batch of rays in one training step for image- or patch-level losses.
arXiv Detail & Related papers (2022-07-10T13:47:20Z) - Multitask AET with Orthogonal Tangent Regularity for Dark Object
Detection [84.52197307286681]
We propose a novel multitask auto encoding transformation (MAET) model to enhance object detection in a dark environment.
In a self-supervision manner, the MAET learns the intrinsic visual structure by encoding and decoding the realistic illumination-degrading transformation.
We have achieved the state-of-the-art performance using synthetic and real-world datasets.
arXiv Detail & Related papers (2022-05-06T16:27:14Z) - PVSeRF: Joint Pixel-, Voxel- and Surface-Aligned Radiance Field for
Single-Image Novel View Synthesis [52.546998369121354]
We present PVSeRF, a learning framework that reconstructs neural radiance fields from single-view RGB images.
We propose to incorporate explicit geometry reasoning and combine it with pixel-aligned features for radiance field prediction.
We show that the introduction of such geometry-aware features helps to achieve a better disentanglement between appearance and geometry.
arXiv Detail & Related papers (2022-02-10T07:39:47Z) - Object-based Illumination Estimation with Rendering-aware Neural
Networks [56.01734918693844]
We present a scheme for fast environment light estimation from the RGBD appearance of individual objects and their local image areas.
With the estimated lighting, virtual objects can be rendered in AR scenarios with shading that is consistent to the real scene.
arXiv Detail & Related papers (2020-08-06T08:23:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.