Don't Forget The Past: Recurrent Depth Estimation from Monocular Video
- URL: http://arxiv.org/abs/2001.02613v2
- Date: Tue, 28 Jul 2020 10:27:13 GMT
- Title: Don't Forget The Past: Recurrent Depth Estimation from Monocular Video
- Authors: Vaishakh Patil, Wouter Van Gansbeke, Dengxin Dai, Luc Van Gool
- Abstract summary: We put three different types of depth estimation into a common framework.
Our method produces a time series of depth maps.
It can be applied to monocular videos only or be combined with different types of sparse depth patterns.
- Score: 92.84498980104424
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autonomous cars need continuously updated depth information. Thus far, depth
is mostly estimated independently for a single frame at a time, even if the
method starts from video input. Our method produces a time series of depth
maps, which makes it an ideal candidate for online learning approaches. In
particular, we put three different types of depth estimation (supervised depth
prediction, self-supervised depth prediction, and self-supervised depth
completion) into a common framework. We integrate the corresponding networks
with a ConvLSTM such that the spatiotemporal structures of depth across frames
can be exploited to yield a more accurate depth estimation. Our method is
flexible. It can be applied to monocular videos only or be combined with
different types of sparse depth patterns. We carefully study the architecture
of the recurrent network and its training strategy. We are first to
successfully exploit recurrent networks for real-time self-supervised monocular
depth estimation and completion. Extensive experiments show that our recurrent
method outperforms its image-based counterpart consistently and significantly
in both self-supervised scenarios. It also outperforms previous depth
estimation methods of the three popular groups. Please refer to
https://www.trace.ethz.ch/publications/2020/rec_depth_estimation/ for details.
Related papers
- Learning Temporally Consistent Video Depth from Video Diffusion Priors [57.929828486615605]
This work addresses the challenge of video depth estimation.
We reformulate the prediction task into a conditional generation problem.
This allows us to leverage the prior knowledge embedded in existing video generation models.
arXiv Detail & Related papers (2024-06-03T16:20:24Z) - Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation [20.230238670888454]
We introduce Marigold, a method for affine-invariant monocular depth estimation.
It can be fine-tuned in a couple of days on a single GPU using only synthetic training data.
It delivers state-of-the-art performance across a wide range of datasets, including over 20% performance gains in specific cases.
arXiv Detail & Related papers (2023-12-04T18:59:13Z) - SC-DepthV3: Robust Self-supervised Monocular Depth Estimation for
Dynamic Scenes [58.89295356901823]
Self-supervised monocular depth estimation has shown impressive results in static scenes.
It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions.
We introduce an external pretrained monocular depth estimation model for generating single-image depth prior.
Our model can predict sharp and accurate depth maps, even when training from monocular videos of highly-dynamic scenes.
arXiv Detail & Related papers (2022-11-07T16:17:47Z) - Towards Accurate Reconstruction of 3D Scene Shape from A Single
Monocular Image [91.71077190961688]
We propose a two-stage framework that first predicts depth up to an unknown scale and shift from a single monocular image.
We then exploits 3D point cloud data to predict the depth shift and the camera's focal length that allow us to recover 3D scene shapes.
We test our depth model on nine unseen datasets and achieve state-of-the-art performance on zero-shot evaluation.
arXiv Detail & Related papers (2022-08-28T16:20:14Z) - How Far Can I Go ? : A Self-Supervised Approach for Deterministic Video
Depth Forecasting [23.134156184783357]
We present a novel self-supervised method to anticipate the depth estimate for a future, unobserved real-world urban scene.
This work is the first to explore self-supervised learning for estimation of monocular depth of future unobserved frames of a video.
arXiv Detail & Related papers (2022-07-01T15:51:17Z) - SelfTune: Metrically Scaled Monocular Depth Estimation through
Self-Supervised Learning [53.78813049373321]
We propose a self-supervised learning method for the pre-trained supervised monocular depth networks to enable metrically scaled depth estimation.
Our approach is useful for various applications such as mobile robot navigation and is applicable to diverse environments.
arXiv Detail & Related papers (2022-03-10T12:28:42Z) - Multi-view Depth Estimation using Epipolar Spatio-Temporal Networks [87.50632573601283]
We present a novel method for multi-view depth estimation from a single video.
Our method achieves temporally coherent depth estimation results by using a novel Epipolar Spatio-Temporal (EST) transformer.
To reduce the computational cost, inspired by recent Mixture-of-Experts models, we design a compact hybrid network.
arXiv Detail & Related papers (2020-11-26T04:04:21Z) - RealMonoDepth: Self-Supervised Monocular Depth Estimation for General
Scenes [11.995578248462946]
Existing supervised methods for monocular depth estimation require accurate depth measurements for training.
Self-supervised approaches have demonstrated impressive results but do not generalise to scenes with different depth ranges or camera baselines.
We introduce RealMonoDepth, a self-supervised monocular depth estimation approach which learns to estimate the real scene depth for a diverse range of indoor and outdoor scenes.
arXiv Detail & Related papers (2020-04-14T02:03:10Z) - Monocular Depth Estimation Based On Deep Learning: An Overview [16.2543991384566]
Inferring depth information from a single image (monocular depth estimation) is an ill-posed problem.
Deep learning has been widely studied recently and achieved promising performance in accuracy.
In order to improve the accuracy of depth estimation, different kinds of network frameworks, loss functions and training strategies are proposed.
arXiv Detail & Related papers (2020-03-14T12:35:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.