M^3VSNet: Unsupervised Multi-metric Multi-view Stereo Network
- URL: http://arxiv.org/abs/2005.00363v2
- Date: Sat, 6 Jun 2020 03:07:12 GMT
- Title: M^3VSNet: Unsupervised Multi-metric Multi-view Stereo Network
- Authors: Baichuan Huang, Hongwei Yi, Can Huang, Yijia He, Jingbin Liu, Xiao Liu
- Abstract summary: We propose a novel unsupervised multi-metric MVS network, named M3VSNet, for dense point cloud reconstruction without supervision.
To improve the robustness and completeness of point cloud reconstruction, we propose a novel multi-metric loss function that combines pixel-wise and feature-wise loss function.
Experimental results show that M3VSNet establishes the state-of-the-arts unsupervised method and achieves comparable performance with previous supervised MVSNet.
- Score: 13.447649324253572
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The present Multi-view stereo (MVS) methods with supervised learning-based
networks have an impressive performance comparing with traditional MVS methods.
However, the ground-truth depth maps for training are hard to be obtained and
are within limited kinds of scenarios. In this paper, we propose a novel
unsupervised multi-metric MVS network, named M^3VSNet, for dense point cloud
reconstruction without any supervision. To improve the robustness and
completeness of point cloud reconstruction, we propose a novel multi-metric
loss function that combines pixel-wise and feature-wise loss function to learn
the inherent constraints from different perspectives of matching
correspondences. Besides, we also incorporate the normal-depth consistency in
the 3D point cloud format to improve the accuracy and continuity of the
estimated depth maps. Experimental results show that M3VSNet establishes the
state-of-the-arts unsupervised method and achieves comparable performance with
previous supervised MVSNet on the DTU dataset and demonstrates the powerful
generalization ability on the Tanks and Temples benchmark with effective
improvement. Our code is available at https://github.com/whubaichuan/M3VSNet
Related papers
- 3DVNet: Multi-View Depth Prediction and Volumetric Refinement [68.68537312256144]
3DVNet is a novel multi-view stereo (MVS) depth-prediction method.
Our key idea is the use of a 3D scene-modeling network that iteratively updates a set of coarse depth predictions.
We show that our method exceeds state-of-the-art accuracy in both depth prediction and 3D reconstruction metrics.
arXiv Detail & Related papers (2021-12-01T00:52:42Z) - TANDEM: Tracking and Dense Mapping in Real-time using Deep Multi-view
Stereo [55.30992853477754]
We present TANDEM, a real-time monocular tracking and dense framework.
For pose estimation, TANDEM performs photometric bundle adjustment based on a sliding window of alignments.
TANDEM shows state-of-the-art real-time 3D reconstruction performance.
arXiv Detail & Related papers (2021-11-14T19:01:02Z) - Digging into Uncertainty in Self-supervised Multi-view Stereo [57.04768354383339]
We propose a novel Uncertainty reduction Multi-view Stereo (UMVS) framework for self-supervised learning.
Our framework achieves the best performance among unsupervised MVS methods, with competitive performance with its supervised opponents.
arXiv Detail & Related papers (2021-08-30T02:53:08Z) - 3D Point Cloud Registration with Multi-Scale Architecture and
Self-supervised Fine-tuning [5.629161809575013]
MS-SVConv is a fast multi-scale deep neural network that outputs features from point clouds for 3D registration between two scenes.
We show significant improvements compared to state-of-the-art methods on the competitive and well-known 3DMatch benchmark.
We present a strategy to fine-tune MS-SVConv on unknown datasets in a self-supervised way, which leads to state-of-the-art results on ETH and TUM datasets.
arXiv Detail & Related papers (2021-03-26T15:38:33Z) - Multi-view Depth Estimation using Epipolar Spatio-Temporal Networks [87.50632573601283]
We present a novel method for multi-view depth estimation from a single video.
Our method achieves temporally coherent depth estimation results by using a novel Epipolar Spatio-Temporal (EST) transformer.
To reduce the computational cost, inspired by recent Mixture-of-Experts models, we design a compact hybrid network.
arXiv Detail & Related papers (2020-11-26T04:04:21Z) - Dense Hybrid Recurrent Multi-view Stereo Net with Dynamic Consistency
Checking [54.58791377183574]
Our novel hybrid recurrent multi-view stereo net consists of two core modules: 1) a light DRENet (Dense Reception Expanded) module to extract dense feature maps of original size with multi-scale context information, 2) a HU-LSTM (Hybrid U-LSTM) to regularize 3D matching volume into predicted depth map.
Our method exhibits competitive performance to the state-of-the-art method while dramatically reduces memory consumption, which costs only $19.4%$ of R-MVSNet memory consumption.
arXiv Detail & Related papers (2020-07-21T14:59:59Z) - SurfaceNet+: An End-to-end 3D Neural Network for Very Sparse Multi-view
Stereopsis [52.35697180864202]
Multi-view stereopsis (MVS) tries to recover the 3D model from 2D images.
We investigate sparse-MVS with large baseline angles since the sparser sensation is more practical and more cost-efficient.
We present SurfaceNet+, a volumetric method to handle the 'incompleteness' and the 'inaccuracy' problems induced by a very sparse MVS setup.
arXiv Detail & Related papers (2020-05-26T13:13:02Z) - M^3VSNet: Unsupervised Multi-metric Multi-view Stereo Network [13.447649324253572]
We propose a novel unsupervised multi-metric MVS network, named M3VSNet, for dense point cloud reconstruction without supervision.
To improve the robustness and completeness of point cloud reconstruction, we propose a novel multi-metric loss function that combines pixel-wise and feature-wise loss function.
Experimental results show that M3VSNet establishes the state-of-the-arts unsupervised method and achieves comparable performance with previous supervised MVSNet.
arXiv Detail & Related papers (2020-04-21T02:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.