RoutedFusion: Learning Real-time Depth Map Fusion
- URL: http://arxiv.org/abs/2001.04388v2
- Date: Fri, 3 Apr 2020 09:15:29 GMT
- Title: RoutedFusion: Learning Real-time Depth Map Fusion
- Authors: Silvan Weder, Johannes L. Sch\"onberger, Marc Pollefeys, Martin R.
Oswald
- Abstract summary: We present a novel real-time capable machine learning-based method for depth map fusion.
We propose a neural network that predicts non-linear updates to better account for typical fusion errors.
Our network is composed of a 2D depth routing network and a 3D depth fusion network which efficiently handle sensor-specific noise and outliers.
- Score: 73.0378509030908
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The efficient fusion of depth maps is a key part of most state-of-the-art 3D
reconstruction methods. Besides requiring high accuracy, these depth fusion
methods need to be scalable and real-time capable. To this end, we present a
novel real-time capable machine learning-based method for depth map fusion.
Similar to the seminal depth map fusion approach by Curless and Levoy, we only
update a local group of voxels to ensure real-time capability. Instead of a
simple linear fusion of depth information, we propose a neural network that
predicts non-linear updates to better account for typical fusion errors. Our
network is composed of a 2D depth routing network and a 3D depth fusion network
which efficiently handle sensor-specific noise and outliers. This is especially
useful for surface edges and thin objects for which the original approach
suffers from thickening artifacts. Our method outperforms the traditional
fusion approach and related learned approaches on both synthetic and real data.
We demonstrate the performance of our method in reconstructing fine geometric
details from noise and outlier contaminated data on various scenes.
Related papers
- Multi-resolution Monocular Depth Map Fusion by Self-supervised
Gradient-based Composition [14.246972408737987]
We propose a novel depth map fusion module to combine the advantages of estimations with multi-resolution inputs.
Our lightweight depth fusion is one-shot and runs in real-time, making our method 80X faster than a state-of-the-art depth fusion method.
arXiv Detail & Related papers (2022-12-03T05:13:50Z) - VolumeFusion: Deep Depth Fusion for 3D Scene Reconstruction [71.83308989022635]
In this paper, we advocate that replicating the traditional two stages framework with deep neural networks improves both the interpretability and the accuracy of the results.
Our network operates in two steps: 1) the local computation of the local depth maps with a deep MVS technique, and, 2) the depth maps and images' features fusion to build a single TSDF volume.
In order to improve the matching performance between images acquired from very different viewpoints, we introduce a rotation-invariant 3D convolution kernel called PosedConv.
arXiv Detail & Related papers (2021-08-19T11:33:58Z) - A Real-Time Online Learning Framework for Joint 3D Reconstruction and
Semantic Segmentation of Indoor Scenes [87.74952229507096]
This paper presents a real-time online vision framework to jointly recover an indoor scene's 3D structure and semantic label.
Given noisy depth maps, a camera trajectory, and 2D semantic labels at train time, the proposed neural network learns to fuse the depth over frames with suitable semantic labels in the scene space.
arXiv Detail & Related papers (2021-08-11T14:29:01Z) - Volumetric Propagation Network: Stereo-LiDAR Fusion for Long-Range Depth
Estimation [81.08111209632501]
We propose a geometry-aware stereo-LiDAR fusion network for long-range depth estimation.
We exploit sparse and accurate point clouds as a cue for guiding correspondences of stereo images in a unified 3D volume space.
Our network achieves state-of-the-art performance on the KITTI and the Virtual- KITTI datasets.
arXiv Detail & Related papers (2021-03-24T03:24:46Z) - NeuralFusion: Online Depth Fusion in Latent Space [77.59420353185355]
We present a novel online depth map fusion approach that learns depth map aggregation in a latent feature space.
Our approach is real-time capable, handles high noise levels, and is particularly able to deal with gross outliers common for photometric stereo-based depth maps.
arXiv Detail & Related papers (2020-11-30T13:50:59Z) - A Single Stream Network for Robust and Real-time RGB-D Salient Object
Detection [89.88222217065858]
We design a single stream network to use the depth map to guide early fusion and middle fusion between RGB and depth.
This model is 55.5% lighter than the current lightest model and runs at a real-time speed of 32 FPS when processing a $384 times 384$ image.
arXiv Detail & Related papers (2020-07-14T04:40:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.