Depth Estimation by Combining Binocular Stereo and Monocular
Structured-Light
- URL: http://arxiv.org/abs/2203.10493v1
- Date: Sun, 20 Mar 2022 08:46:37 GMT
- Title: Depth Estimation by Combining Binocular Stereo and Monocular
Structured-Light
- Authors: Yuhua Xu, Xiaoli Yang, Yushan Yu, Wei Jia, Zhaobi Chu, Yulan Guo
- Abstract summary: We present a novel stereo system, which consists of two cameras (an RGB camera and an IR camera) and an IR speckle projector.
The RGB camera is used both for depth estimation and texture acquisition.
The depth map generated by the MSL subsystem can provide external guidance for the stereo matching networks.
- Score: 29.226203202113613
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is well known that the passive stereo system cannot adapt well to weak
texture objects, e.g., white walls. However, these weak texture targets are
very common in indoor environments. In this paper, we present a novel stereo
system, which consists of two cameras (an RGB camera and an IR camera) and an
IR speckle projector. The RGB camera is used both for depth estimation and
texture acquisition. The IR camera and the speckle projector can form a
monocular structured-light (MSL) subsystem, while the two cameras can form a
binocular stereo subsystem. The depth map generated by the MSL subsystem can
provide external guidance for the stereo matching networks, which can improve
the matching accuracy significantly. In order to verify the effectiveness of
the proposed system, we build a prototype and collect a test dataset in indoor
scenes. The evaluation results show that the Bad 2.0 error of the proposed
system is 28.2% of the passive stereo system when the network RAFT is used. The
dataset and trained models are available at
https://github.com/YuhuaXu/MonoStereoFusion.
Related papers
- ASGrasp: Generalizable Transparent Object Reconstruction and Grasping from RGB-D Active Stereo Camera [9.212504138203222]
We propose ASGrasp, a 6-DoF grasp detection network that uses an RGB-D active stereo camera.
Our system distinguishes itself by its ability to directly utilize raw IR and RGB images for transparent object geometry reconstruction.
Our experiments demonstrate that ASGrasp can achieve over 90% success rate for generalizable transparent object grasping.
arXiv Detail & Related papers (2024-05-09T09:44:51Z) - SDGE: Stereo Guided Depth Estimation for 360$^\circ$ Camera Sets [65.64958606221069]
Multi-camera systems are often used in autonomous driving to achieve a 360$circ$ perception.
These 360$circ$ camera sets often have limited or low-quality overlap regions, making multi-view stereo methods infeasible for the entire image.
We propose the Stereo Guided Depth Estimation (SGDE) method, which enhances depth estimation of the full image by explicitly utilizing multi-view stereo results on the overlap.
arXiv Detail & Related papers (2024-02-19T02:41:37Z) - Unsupervised Visible-light Images Guided Cross-Spectrum Depth Estimation
from Dual-Modality Cameras [33.77748026254935]
Cross-spectrum depth estimation aims to provide a depth map in all illumination conditions with a pair of dual-spectrum images.
In this paper, we propose an unsupervised visible-light image guided cross-spectrum (i.e., thermal and visible-light, TIR-VIS in short) depth estimation framework.
Our method achieves better performance than the compared existing methods.
arXiv Detail & Related papers (2022-04-30T12:58:35Z) - Self-Supervised Depth Completion for Active Stereo [55.79929735390945]
Active stereo systems are widely used in the robotics industry due to their low cost and high quality depth maps.
These depth sensors suffer from stereo artefacts and do not provide dense depth estimates.
We present the first self-supervised depth completion method for active stereo systems that predicts accurate dense depth maps.
arXiv Detail & Related papers (2021-10-07T07:33:52Z) - MFuseNet: Robust Depth Estimation with Learned Multiscopic Fusion [47.2251122861135]
We design a multiscopic vision system that utilizes a low-cost monocular RGB camera to acquire accurate depth estimation.
Unlike multi-view stereo with images captured at unconstrained camera poses, the proposed system controls the motion of a camera to capture a sequence of images.
arXiv Detail & Related papers (2021-08-05T08:31:01Z) - High-Resolution Depth Maps Based on TOF-Stereo Fusion [27.10059147107254]
We propose a novel TOF-stereo fusion method based on an efficient seed-growing algorithm.
We show that the proposed algorithm outperforms 2D image-based stereo algorithms.
The algorithm potentially exhibits real-time performance on a single CPU.
arXiv Detail & Related papers (2021-07-30T15:11:42Z) - OmniSLAM: Omnidirectional Localization and Dense Mapping for
Wide-baseline Multi-camera Systems [88.41004332322788]
We present an omnidirectional localization and dense mapping system for a wide-baseline multiview stereo setup with ultra-wide field-of-view (FOV) fisheye cameras.
For more practical and accurate reconstruction, we first introduce improved and light-weighted deep neural networks for the omnidirectional depth estimation.
We integrate our omnidirectional depth estimates into the visual odometry (VO) and add a loop closing module for global consistency.
arXiv Detail & Related papers (2020-03-18T05:52:10Z) - Active Perception with A Monocular Camera for Multiscopic Vision [50.370074098619185]
We design a multiscopic vision system that utilizes a low-cost monocular RGB camera to acquire accurate depth estimation for robotic applications.
Unlike multi-view stereo with images captured at unconstrained camera poses, the proposed system actively controls a robot arm with a mounted camera to capture a sequence of images in horizontally or vertically aligned positions with the same parallax.
arXiv Detail & Related papers (2020-01-22T08:46:45Z) - Multi-View Photometric Stereo: A Robust Solution and Benchmark Dataset
for Spatially Varying Isotropic Materials [65.95928593628128]
We present a method to capture both 3D shape and spatially varying reflectance with a multi-view photometric stereo technique.
Our algorithm is suitable for perspective cameras and nearby point light sources.
arXiv Detail & Related papers (2020-01-18T12:26:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.