Self-Attention Dense Depth Estimation Network for Unrectified Video
Sequences
- URL: http://arxiv.org/abs/2005.14313v1
- Date: Thu, 28 May 2020 21:53:53 GMT
- Title: Self-Attention Dense Depth Estimation Network for Unrectified Video
Sequences
- Authors: Alwyn Mathew, Aditya Prakash Patra, Jimson Mathew
- Abstract summary: LiDAR and radar sensors are the hardware solution for real-time depth estimation.
Deep learning based self-supervised depth estimation methods have shown promising results.
We propose a self-attention based depth and ego-motion network for unrectified images.
- Score: 6.821598757786515
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The dense depth estimation of a 3D scene has numerous applications, mainly in
robotics and surveillance. LiDAR and radar sensors are the hardware solution
for real-time depth estimation, but these sensors produce sparse depth maps and
are sometimes unreliable. In recent years research aimed at tackling depth
estimation using single 2D image has received a lot of attention. The deep
learning based self-supervised depth estimation methods from the rectified
stereo and monocular video frames have shown promising results. We propose a
self-attention based depth and ego-motion network for unrectified images. We
also introduce non-differentiable distortion of the camera into the training
pipeline. Our approach performs competitively when compared to other
established approaches that used rectified images for depth estimation.
Related papers
- SDGE: Stereo Guided Depth Estimation for 360$^\circ$ Camera Sets [65.64958606221069]
Multi-camera systems are often used in autonomous driving to achieve a 360$circ$ perception.
These 360$circ$ camera sets often have limited or low-quality overlap regions, making multi-view stereo methods infeasible for the entire image.
We propose the Stereo Guided Depth Estimation (SGDE) method, which enhances depth estimation of the full image by explicitly utilizing multi-view stereo results on the overlap.
arXiv Detail & Related papers (2024-02-19T02:41:37Z) - Blur aware metric depth estimation with multi-focus plenoptic cameras [8.508198765617196]
We present a new metric depth estimation algorithm using only raw images from a multi-focus plenoptic camera.
The proposed approach is especially suited for the multi-focus configuration where several micro-lenses with different focal lengths are used.
arXiv Detail & Related papers (2023-08-08T13:38:50Z) - SC-DepthV3: Robust Self-supervised Monocular Depth Estimation for
Dynamic Scenes [58.89295356901823]
Self-supervised monocular depth estimation has shown impressive results in static scenes.
It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions.
We introduce an external pretrained monocular depth estimation model for generating single-image depth prior.
Our model can predict sharp and accurate depth maps, even when training from monocular videos of highly-dynamic scenes.
arXiv Detail & Related papers (2022-11-07T16:17:47Z) - Unsupervised confidence for LiDAR depth maps and applications [43.474845978673166]
We propose an effective unsupervised framework aimed at addressing the issue of sparse LiDAR depth maps.
Our framework estimates the confidence of the sparse depth map and thus allows for filtering out the outliers.
We demonstrate how this achievement can improve a wide range of tasks.
arXiv Detail & Related papers (2022-10-06T17:59:58Z) - BEVDepth: Acquisition of Reliable Depth for Multi-view 3D Object
Detection [13.319949358652192]
We propose a new 3D object detector with a trustworthy depth estimation, dubbed BEVDepth, for camera-based Bird's-Eye-View 3D object detection.
BEVDepth achieves the new state-of-the-art 60.0% NDS on the challenging nuScenes test set.
arXiv Detail & Related papers (2022-06-21T03:21:18Z) - SurroundDepth: Entangling Surrounding Views for Self-Supervised
Multi-Camera Depth Estimation [101.55622133406446]
We propose a SurroundDepth method to incorporate the information from multiple surrounding views to predict depth maps across cameras.
Specifically, we employ a joint network to process all the surrounding views and propose a cross-view transformer to effectively fuse the information from multiple views.
In experiments, our method achieves the state-of-the-art performance on the challenging multi-camera depth estimation datasets.
arXiv Detail & Related papers (2022-04-07T17:58:47Z) - Learning Occlusion-Aware Coarse-to-Fine Depth Map for Self-supervised
Monocular Depth Estimation [11.929584800629673]
We propose a novel network to learn an Occlusion-aware Coarse-to-Fine Depth map for self-supervised monocular depth estimation.
The proposed OCFD-Net does not only employ a discrete depth constraint for learning a coarse-level depth map, but also employ a continuous depth constraint for learning a scene depth residual.
arXiv Detail & Related papers (2022-03-21T12:43:42Z) - Gated2Gated: Self-Supervised Depth Estimation from Gated Images [22.415893281441928]
Gated cameras hold promise as an alternative to scanning LiDAR sensors with high-resolution 3D depth.
We propose an entirely self-supervised depth estimation method that uses gated intensity profiles and temporal consistency as a training signal.
arXiv Detail & Related papers (2021-12-04T19:47:38Z) - DnD: Dense Depth Estimation in Crowded Dynamic Indoor Scenes [68.38952377590499]
We present a novel approach for estimating depth from a monocular camera as it moves through complex indoor environments.
Our approach predicts absolute scale depth maps over the entire scene consisting of a static background and multiple moving people.
arXiv Detail & Related papers (2021-08-12T09:12:39Z) - Don't Forget The Past: Recurrent Depth Estimation from Monocular Video [92.84498980104424]
We put three different types of depth estimation into a common framework.
Our method produces a time series of depth maps.
It can be applied to monocular videos only or be combined with different types of sparse depth patterns.
arXiv Detail & Related papers (2020-01-08T16:50:51Z) - Video Depth Estimation by Fusing Flow-to-Depth Proposals [65.24533384679657]
We present an approach with a differentiable flow-to-depth layer for video depth estimation.
The model consists of a flow-to-depth layer, a camera pose refinement module, and a depth fusion network.
Our approach outperforms state-of-the-art depth estimation methods, and has reasonable cross dataset generalization capability.
arXiv Detail & Related papers (2019-12-30T10:45:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.