Self-Supervised Depth Completion for Active Stereo
- URL: http://arxiv.org/abs/2110.03234v1
- Date: Thu, 7 Oct 2021 07:33:52 GMT
- Title: Self-Supervised Depth Completion for Active Stereo
- Authors: Frederik Warburg, Daniel Hernandez-Juarez, Juan Tarrio, Alexander
Vakhitov, Ujwal Bonde, Pablo Alcantarilla
- Abstract summary: Active stereo systems are widely used in the robotics industry due to their low cost and high quality depth maps.
These depth sensors suffer from stereo artefacts and do not provide dense depth estimates.
We present the first self-supervised depth completion method for active stereo systems that predicts accurate dense depth maps.
- Score: 55.79929735390945
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Active stereo systems are widely used in the robotics industry due to their
low cost and high quality depth maps. These depth sensors, however, suffer from
stereo artefacts and do not provide dense depth estimates. In this work, we
present the first self-supervised depth completion method for active stereo
systems that predicts accurate dense depth maps. Our system leverages a
feature-based visual inertial SLAM system to produce motion estimates and
accurate (but sparse) 3D landmarks. The 3D landmarks are used both as model
input and as supervision during training. The motion estimates are used in our
novel reconstruction loss that relies on a combination of passive and active
stereo frames, resulting in significant improvements in textureless areas that
are common in indoor environments. Due to the non-existence of publicly
available active stereo datasets, we release a real dataset together with
additional information for a publicly available synthetic dataset needed for
active depth completion and prediction. Through rigorous evaluations we show
that our method outperforms state of the art on both datasets. Additionally we
show how our method obtains more complete, and therefore safer, 3D maps when
used in a robotic platform
Related papers
- Stereo-Depth Fusion through Virtual Pattern Projection [37.519762078762575]
This paper presents a novel general-purpose stereo and depth data fusion paradigm.
It mimics the active stereo principle by replacing the unreliable physical pattern projector with a depth sensor.
It works by projecting virtual patterns consistent with the scene geometry onto the left and right images acquired by a conventional stereo camera.
arXiv Detail & Related papers (2024-06-06T17:59:58Z) - DynamicStereo: Consistent Dynamic Depth from Stereo Videos [91.1804971397608]
We propose DynamicStereo to estimate disparity for stereo videos.
The network learns to pool information from neighboring frames to improve the temporal consistency of its predictions.
We also introduce Dynamic Replica, a new benchmark dataset containing synthetic videos of people and animals in scanned environments.
arXiv Detail & Related papers (2023-05-03T17:40:49Z) - Semi-Perspective Decoupled Heatmaps for 3D Robot Pose Estimation from
Depth Maps [66.24554680709417]
Knowing the exact 3D location of workers and robots in a collaborative environment enables several real applications.
We propose a non-invasive framework based on depth devices and deep neural networks to estimate the 3D pose of robots from an external camera.
arXiv Detail & Related papers (2022-07-06T08:52:12Z) - DSGN++: Exploiting Visual-Spatial Relation forStereo-based 3D Detectors [60.88824519770208]
Camera-based 3D object detectors are welcome due to their wider deployment and lower price than LiDAR sensors.
We revisit the prior stereo modeling DSGN about the stereo volume constructions for representing both 3D geometry and semantics.
We propose our approach, DSGN++, aiming for improving information flow throughout the 2D-to-3D pipeline.
arXiv Detail & Related papers (2022-04-06T18:43:54Z) - 3DVNet: Multi-View Depth Prediction and Volumetric Refinement [68.68537312256144]
3DVNet is a novel multi-view stereo (MVS) depth-prediction method.
Our key idea is the use of a 3D scene-modeling network that iteratively updates a set of coarse depth predictions.
We show that our method exceeds state-of-the-art accuracy in both depth prediction and 3D reconstruction metrics.
arXiv Detail & Related papers (2021-12-01T00:52:42Z) - A Learned Stereo Depth System for Robotic Manipulation in Homes [2.06216858680643]
We present a passive stereo depth system that produces dense and accurate point clouds optimized for human environments.
The system consists of an algorithm combining learned stereo matching with engineered filtering, a training and data-mixing methodology, and a sensor hardware design.
arXiv Detail & Related papers (2021-09-23T20:53:55Z) - SIDE: Center-based Stereo 3D Detector with Structure-aware Instance
Depth Estimation [11.169586369931803]
We propose a stereo-image based anchor-free 3D detection method, called structure-aware stereo 3D detector (termed as SIDE)
We explore the instance-level depth information via constructing the cost volume from RoIs of each object.
Our method achieves the state-of-the-art performance compared to existing methods without depth map supervision.
arXiv Detail & Related papers (2021-08-22T08:09:38Z) - OmniSLAM: Omnidirectional Localization and Dense Mapping for
Wide-baseline Multi-camera Systems [88.41004332322788]
We present an omnidirectional localization and dense mapping system for a wide-baseline multiview stereo setup with ultra-wide field-of-view (FOV) fisheye cameras.
For more practical and accurate reconstruction, we first introduce improved and light-weighted deep neural networks for the omnidirectional depth estimation.
We integrate our omnidirectional depth estimates into the visual odometry (VO) and add a loop closing module for global consistency.
arXiv Detail & Related papers (2020-03-18T05:52:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.