WGANVO: Monocular Visual Odometry based on Generative Adversarial
Networks
- URL: http://arxiv.org/abs/2007.13704v1
- Date: Mon, 27 Jul 2020 17:31:24 GMT
- Title: WGANVO: Monocular Visual Odometry based on Generative Adversarial
Networks
- Authors: Javier Cremona, Lucas Uzal, Taih\'u Pire
- Abstract summary: WGANVO is a Deep Learning based monocular Visual Odometry method.
In particular, a neural network is trained to regress a pose estimate from an image pair.
The method can recover the absolute scale of the scene without neither prior knowledge nor extra information.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we present WGANVO, a Deep Learning based monocular Visual
Odometry method. In particular, a neural network is trained to regress a pose
estimate from an image pair. The training is performed using a semi-supervised
approach. Unlike geometry based monocular methods, the proposed method can
recover the absolute scale of the scene without neither prior knowledge nor
extra information. The evaluation of the system is carried out on the
well-known KITTI dataset where it is shown to work in real time and the
accuracy obtained is encouraging to continue the development of Deep Learning
based methods.
Related papers
- Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting & Beyond [61.18736646013446]
In pursuit of a deeper understanding of its surprising behaviors, we investigate the utility of a simple yet accurate model of a trained neural network.
Across three case studies, we illustrate how it can be applied to derive new empirical insights on a diverse range of prominent phenomena.
arXiv Detail & Related papers (2024-10-31T22:54:34Z) - ALSO: Automotive Lidar Self-supervision by Occupancy estimation [70.70557577874155]
We propose a new self-supervised method for pre-training the backbone of deep perception models operating on point clouds.
The core idea is to train the model on a pretext task which is the reconstruction of the surface on which the 3D points are sampled.
The intuition is that if the network is able to reconstruct the scene surface, given only sparse input points, then it probably also captures some fragments of semantic information.
arXiv Detail & Related papers (2022-12-12T13:10:19Z) - Neural Maximum A Posteriori Estimation on Unpaired Data for Motion
Deblurring [87.97330195531029]
We propose a Neural Maximum A Posteriori (NeurMAP) estimation framework for training neural networks to recover blind motion information and sharp content from unpaired data.
The proposed NeurMAP is an approach to existing deblurring neural networks, and is the first framework that enables training image deblurring networks on unpaired datasets.
arXiv Detail & Related papers (2022-04-26T08:09:47Z) - SelfTune: Metrically Scaled Monocular Depth Estimation through
Self-Supervised Learning [53.78813049373321]
We propose a self-supervised learning method for the pre-trained supervised monocular depth networks to enable metrically scaled depth estimation.
Our approach is useful for various applications such as mobile robot navigation and is applicable to diverse environments.
arXiv Detail & Related papers (2022-03-10T12:28:42Z) - X-Distill: Improving Self-Supervised Monocular Depth via Cross-Task
Distillation [69.9604394044652]
We propose a novel method to improve the self-supervised training of monocular depth via cross-task knowledge distillation.
During training, we utilize a pretrained semantic segmentation teacher network and transfer its semantic knowledge to the depth network.
We extensively evaluate the efficacy of our proposed approach on the KITTI benchmark and compare it with the latest state of the art.
arXiv Detail & Related papers (2021-10-24T19:47:14Z) - Self-Supervised Monocular Depth Estimation with Internal Feature Fusion [12.874712571149725]
Self-supervised learning for depth estimation uses geometry in image sequences for supervision.
We propose a novel depth estimation networkDIFFNet, which can make use of semantic information in down and upsampling procedures.
arXiv Detail & Related papers (2021-10-18T17:31:11Z) - Generalizing to the Open World: Deep Visual Odometry with Online
Adaptation [27.22639812204019]
We propose an online adaptation framework for deep VO with the assistance of scene-agnostic geometric computations and Bayesian inference.
Our method achieves state-of-the-art generalization ability among self-supervised VO methods.
arXiv Detail & Related papers (2021-03-29T02:13:56Z) - Deep Learning based Monocular Depth Prediction: Datasets, Methods and
Applications [31.06326714016336]
Estimating depth from RGB images can facilitate many computer vision tasks, such as indoor localization, height estimation, and simultaneous localization and mapping.
Recently, monocular depth estimation has obtained great progress owing to the rapid development of deep learning techniques.
arXiv Detail & Related papers (2020-11-09T01:03:13Z) - Semantically-Guided Representation Learning for Self-Supervised
Monocular Depth [40.49380547487908]
We propose a new architecture leveraging fixed pretrained semantic segmentation networks to guide self-supervised representation learning.
Our method improves upon the state of the art for self-supervised monocular depth prediction over all pixels, fine-grained details, and per semantic categories.
arXiv Detail & Related papers (2020-02-27T18:40:10Z) - Single Image Depth Estimation Trained via Depth from Defocus Cues [105.67073923825842]
Estimating depth from a single RGB image is a fundamental task in computer vision.
In this work, we rely, instead of different views, on depth from focus cues.
We present results that are on par with supervised methods on KITTI and Make3D datasets and outperform unsupervised learning approaches.
arXiv Detail & Related papers (2020-01-14T20:22:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.