MP-MVS: Multi-Scale Windows PatchMatch and Planar Prior Multi-View
Stereo
- URL: http://arxiv.org/abs/2309.13294v1
- Date: Sat, 23 Sep 2023 07:30:42 GMT
- Title: MP-MVS: Multi-Scale Windows PatchMatch and Planar Prior Multi-View
Stereo
- Authors: Rongxuan Tan, Qing Wang, Xueyan Wang, Chao Yan, Yang Sun and Youyang
Feng
- Abstract summary: We propose a resilient and effective multi-view stereo approach (MP-MVS)
We design a multi-scale windows PatchMatch (mPM) to obtain reliable depth of untextured areas.
In contrast with other multi-scale approaches, which is faster and can be easily extended to PatchMatch-based MVS approaches.
- Score: 7.130834755320434
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Significant strides have been made in enhancing the accuracy of Multi-View
Stereo (MVS)-based 3D reconstruction. However, untextured areas with unstable
photometric consistency often remain incompletely reconstructed. In this paper,
we propose a resilient and effective multi-view stereo approach (MP-MVS). We
design a multi-scale windows PatchMatch (mPM) to obtain reliable depth of
untextured areas. In contrast with other multi-scale approaches, which is
faster and can be easily extended to PatchMatch-based MVS approaches.
Subsequently, we improve the existing checkerboard sampling schemes by limiting
our sampling to distant regions, which can effectively improve the efficiency
of spatial propagation while mitigating outlier generation. Finally, we
introduce and improve planar prior assisted PatchMatch of ACMP. Instead of
relying on photometric consistency, we utilize geometric consistency
information between multi-views to select reliable triangulated vertices. This
strategy can obtain a more accurate planar prior model to rectify photometric
consistency measurements. Our approach has been tested on the ETH3D High-res
multi-view benchmark with several state-of-the-art approaches. The results
demonstrate that our approach can reach the state-of-the-art. The associated
codes will be accessible at https://github.com/RongxuanTan/MP-MVS.
Related papers
- A Global Depth-Range-Free Multi-View Stereo Transformer Network with Pose Embedding [76.44979557843367]
We propose a novel multi-view stereo (MVS) framework that gets rid of the depth range prior.
We introduce a Multi-view Disparity Attention (MDA) module to aggregate long-range context information.
We explicitly estimate the quality of the current pixel corresponding to sampled points on the epipolar line of the source image.
arXiv Detail & Related papers (2024-11-04T08:50:16Z) - SD-MVS: Segmentation-Driven Deformation Multi-View Stereo with Spherical
Refinement and EM optimization [6.886220026399106]
We introduce Multi-View Stereo (SD-MVS) to tackle challenges in 3D reconstruction of textureless areas.
We are the first to adopt the Segment Anything Model (SAM) to distinguish semantic instances in scenes.
We propose a unique refinement strategy that combines spherical coordinates and gradient descent on normals and pixelwise search interval on depths.
arXiv Detail & Related papers (2024-01-12T05:25:57Z) - Deep PatchMatch MVS with Learned Patch Coplanarity, Geometric
Consistency and Adaptive Pixel Sampling [19.412014102866507]
We build on learning-based approaches to improve photometric scores by learning patch coplanarity and encourage geometric consistency.
We propose an adaptive pixel sampling strategy for candidate propagation that reduces memory to enable training on larger resolution with more views and a larger encoder.
arXiv Detail & Related papers (2022-10-14T07:29:03Z) - Multiview Stereo with Cascaded Epipolar RAFT [73.7619703879639]
We address multiview stereo (MVS), an important 3D vision task that reconstructs a 3D model such as a dense point cloud from multiple calibrated images.
We propose CER-MVS, a new approach based on the RAFT (Recurrent All-Pairs Field Transforms) architecture developed for optical flow. CER-MVS introduces five new changes to RAFT: epipolar cost volumes, cost volume cascading, multiview fusion of cost volumes, dynamic supervision, and multiresolution fusion of depth maps.
arXiv Detail & Related papers (2022-05-09T18:17:05Z) - PatchMVSNet: Patch-wise Unsupervised Multi-View Stereo for
Weakly-Textured Surface Reconstruction [2.9896482273918434]
This paper proposes robust loss functions leveraging constraints beneath multi-view images to alleviate matching ambiguity.
Our strategy can be implemented with arbitrary depth estimation frameworks and can be trained with arbitrary large-scale MVS datasets.
Our method reaches the performance of the state-of-the-art methods on popular benchmarks, like DTU, Tanks and Temples and ETH3D.
arXiv Detail & Related papers (2022-03-04T07:05:23Z) - IterMVS: Iterative Probability Estimation for Efficient Multi-View
Stereo [71.84742490020611]
IterMVS is a new data-driven method for high-resolution multi-view stereo.
We propose a novel GRU-based estimator that encodes pixel-wise probability distributions of depth in its hidden state.
We verify the efficiency and effectiveness of our method on DTU, Tanks&Temples and ETH3D.
arXiv Detail & Related papers (2021-12-09T18:58:02Z) - TANDEM: Tracking and Dense Mapping in Real-time using Deep Multi-view
Stereo [55.30992853477754]
We present TANDEM, a real-time monocular tracking and dense framework.
For pose estimation, TANDEM performs photometric bundle adjustment based on a sliding window of alignments.
TANDEM shows state-of-the-art real-time 3D reconstruction performance.
arXiv Detail & Related papers (2021-11-14T19:01:02Z) - Direct Multi-view Multi-person 3D Pose Estimation [138.48139701871213]
We present Multi-view Pose transformer (MvP) for estimating multi-person 3D poses from multi-view images.
MvP directly regresses the multi-person 3D poses in a clean and efficient way, without relying on intermediate tasks.
We show experimentally that our MvP model outperforms the state-of-the-art methods on several benchmarks while being much more efficient.
arXiv Detail & Related papers (2021-11-07T13:09:20Z) - PatchMatch-RL: Deep MVS with Pixelwise Depth, Normal, and Visibility [23.427619869594437]
We propose an end-to-end trainable PatchMatch-based MVS approach that combines advantages of trainable costs and regularizations with pixelwise estimates.
We evaluate our method on widely used MVS benchmarks, ETH3D and Tanks and Temples (TnT)
arXiv Detail & Related papers (2021-08-19T23:14:48Z) - PatchmatchNet: Learned Multi-View Patchmatch Stereo [70.14789588576438]
We present PatchmatchNet, a novel and learnable cascade formulation of Patchmatch for high-resolution multi-view stereo.
With high speed and low memory requirement, PatchmatchNet can process higher resolution imagery and is more suited to run on resource limited devices than competitors that employ 3D cost volume regularization.
arXiv Detail & Related papers (2020-12-02T18:59:02Z) - Planar Prior Assisted PatchMatch Multi-View Stereo [32.41293572426403]
completeness of 3D models is still a challenging problem in multi-view stereo.
Planar models are advantageous to the depth estimation of low-textured areas.
PatchMatch multi-view stereo is very efficient for its sampling and propagation scheme.
arXiv Detail & Related papers (2019-12-26T01:34:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.