MFuseNet: Robust Depth Estimation with Learned Multiscopic Fusion
- URL: http://arxiv.org/abs/2108.02448v2
- Date: Fri, 6 Aug 2021 07:31:12 GMT
- Title: MFuseNet: Robust Depth Estimation with Learned Multiscopic Fusion
- Authors: Weihao Yuan, Rui Fan, Michael Yu Wang, Qifeng Chen
- Abstract summary: We design a multiscopic vision system that utilizes a low-cost monocular RGB camera to acquire accurate depth estimation.
Unlike multi-view stereo with images captured at unconstrained camera poses, the proposed system controls the motion of a camera to capture a sequence of images.
- Score: 47.2251122861135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We design a multiscopic vision system that utilizes a low-cost monocular RGB
camera to acquire accurate depth estimation. Unlike multi-view stereo with
images captured at unconstrained camera poses, the proposed system controls the
motion of a camera to capture a sequence of images in horizontally or
vertically aligned positions with the same parallax. In this system, we propose
a new heuristic method and a robust learning-based method to fuse multiple cost
volumes between the reference image and its surrounding images. To obtain
training data, we build a synthetic dataset with multiscopic images. The
experiments on the real-world Middlebury dataset and real robot demonstration
show that our multiscopic vision system outperforms traditional two-frame
stereo matching methods in depth estimation. Our code and dataset are available
at https://sites.google.com/view/multiscopic.
Related papers
- Learning Robust Multi-Scale Representation for Neural Radiance Fields
from Unposed Images [65.41966114373373]
We present an improved solution to the neural image-based rendering problem in computer vision.
The proposed approach could synthesize a realistic image of the scene from a novel viewpoint at test time.
arXiv Detail & Related papers (2023-11-08T08:18:23Z) - A Flexible Multi-view Multi-modal Imaging System for Outdoor Scenes [37.5716419191546]
We propose a wireless multi-view multi-modal 3D imaging system generally applicable to large outdoor scenes.
Multiple spatially distributed slave nodes equipped with cameras and LiDARs are connected to form a wireless sensor network.
This system is the first imaging system that integrates mutli-view RGB cameras and LiDARs in large outdoor scenes.
arXiv Detail & Related papers (2023-02-21T06:14:05Z) - MEStereo-Du2CNN: A Novel Dual Channel CNN for Learning Robust Depth
Estimates from Multi-exposure Stereo Images for HDR 3D Applications [0.22940141855172028]
We develop a novel deep architecture for multi-exposure stereo depth estimation.
For the stereo depth estimation component of our architecture, a mono-to-stereo transfer learning approach is deployed.
In terms of performance, the proposed model surpasses state-of-the-art monocular and stereo depth estimation methods.
arXiv Detail & Related papers (2022-06-21T13:23:22Z) - SurroundDepth: Entangling Surrounding Views for Self-Supervised
Multi-Camera Depth Estimation [101.55622133406446]
We propose a SurroundDepth method to incorporate the information from multiple surrounding views to predict depth maps across cameras.
Specifically, we employ a joint network to process all the surrounding views and propose a cross-view transformer to effectively fuse the information from multiple views.
In experiments, our method achieves the state-of-the-art performance on the challenging multi-camera depth estimation datasets.
arXiv Detail & Related papers (2022-04-07T17:58:47Z) - MC-Blur: A Comprehensive Benchmark for Image Deblurring [127.6301230023318]
In most real-world images, blur is caused by different factors, e.g., motion and defocus.
We construct a new large-scale multi-cause image deblurring dataset (called MC-Blur)
Based on the MC-Blur dataset, we conduct extensive benchmarking studies to compare SOTA methods in different scenarios.
arXiv Detail & Related papers (2021-12-01T02:10:42Z) - DeepMultiCap: Performance Capture of Multiple Characters Using Sparse
Multiview Cameras [63.186486240525554]
DeepMultiCap is a novel method for multi-person performance capture using sparse multi-view cameras.
Our method can capture time varying surface details without the need of using pre-scanned template models.
arXiv Detail & Related papers (2021-05-01T14:32:13Z) - Active Perception with A Monocular Camera for Multiscopic Vision [50.370074098619185]
We design a multiscopic vision system that utilizes a low-cost monocular RGB camera to acquire accurate depth estimation for robotic applications.
Unlike multi-view stereo with images captured at unconstrained camera poses, the proposed system actively controls a robot arm with a mounted camera to capture a sequence of images in horizontally or vertically aligned positions with the same parallax.
arXiv Detail & Related papers (2020-01-22T08:46:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.