Enhancing Multi-view Stereo with Contrastive Matching and Weighted Focal
Loss
- URL: http://arxiv.org/abs/2206.10360v1
- Date: Tue, 21 Jun 2022 13:10:14 GMT
- Title: Enhancing Multi-view Stereo with Contrastive Matching and Weighted Focal
Loss
- Authors: Yikang Ding, Zhenyang Li, Dihe Huang, Zhiheng Li, Kai Zhang
- Abstract summary: We propose a new method to enhance the performance of existing networks inspired by contrastive learning and feature matching.
Our method achieves state-of-the-art performance and significant improvement over baseline network.
- Score: 10.847120224170698
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning-based multi-view stereo (MVS) methods have made impressive progress
and surpassed traditional methods in recent years. However, their accuracy and
completeness are still struggling. In this paper, we propose a new method to
enhance the performance of existing networks inspired by contrastive learning
and feature matching. First, we propose a Contrast Matching Loss (CML), which
treats the correct matching points in depth-dimension as positive sample and
other points as negative samples, and computes the contrastive loss based on
the similarity of features. We further propose a Weighted Focal Loss (WFL) for
better classification capability, which weakens the contribution of
low-confidence pixels in unimportant areas to the loss according to predicted
confidence. Extensive experiments performed on DTU, Tanks and Temples and
BlendedMVS datasets show our method achieves state-of-the-art performance and
significant improvement over baseline network.
Related papers
- Improving Neural Surface Reconstruction with Feature Priors from Multi-View Image [87.00660347447494]
Recent advancements in Neural Surface Reconstruction (NSR) have significantly improved multi-view reconstruction when coupled with volume rendering.
We propose an investigation into feature-level consistent loss, aiming to harness valuable feature priors from diverse pretext visual tasks.
Our results, analyzed on DTU and EPFL, reveal that feature priors from image matching and multi-view stereo datasets outperform other pretext tasks.
arXiv Detail & Related papers (2024-08-04T16:09:46Z) - Bayesian Learning-driven Prototypical Contrastive Loss for Class-Incremental Learning [42.14439854721613]
We propose a prototypical network with a Bayesian learning-driven contrastive loss (BLCL) tailored specifically for class-incremental learning scenarios.
Our approach dynamically adapts the balance between the cross-entropy and contrastive loss functions with a Bayesian learning technique.
arXiv Detail & Related papers (2024-05-17T19:49:02Z) - Adaptive Learning for Multi-view Stereo Reconstruction [6.635583283522551]
We first analyze existing loss functions' properties for deep depth based MVS approaches.
We then propose a novel loss function, named adaptive Wasserstein loss, which is able to narrow down the difference between the true and predicted probability distributions of depth.
Experiments on different benchmarks, including DTU, Tanks and Temples and BlendedMVS, show that the proposed method with the adaptive Wasserstein loss and the offset module achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-04-08T04:13:35Z) - Contrastive Pre-Training with Multi-View Fusion for No-Reference Point Cloud Quality Assessment [49.36799270585947]
No-reference point cloud quality assessment (NR-PCQA) aims to automatically evaluate the perceptual quality of distorted point clouds without available reference.
We propose a novel contrastive pre-training framework tailored for PCQA (CoPA)
Our method outperforms the state-of-the-art PCQA methods on popular benchmarks.
arXiv Detail & Related papers (2024-03-15T07:16:07Z) - Tuned Contrastive Learning [77.67209954169593]
We propose a novel contrastive loss function -- Tuned Contrastive Learning (TCL) loss.
TCL generalizes to multiple positives and negatives in a batch and offers parameters to tune and improve the gradient responses from hard positives and hard negatives.
We show how to extend TCL to self-supervised setting and empirically compare it with various SOTA self-supervised learning methods.
arXiv Detail & Related papers (2023-05-18T03:26:37Z) - Siamese Prototypical Contrastive Learning [24.794022951873156]
Contrastive Self-supervised Learning (CSL) is a practical solution that learns meaningful visual representations from massive data in an unsupervised approach.
In this paper, we tackle this problem by introducing a simple but effective contrastive learning framework.
The key insight is to employ siamese-style metric loss to match intra-prototype features, while increasing the distance between inter-prototype features.
arXiv Detail & Related papers (2022-08-18T13:25:30Z) - Improving Music Performance Assessment with Contrastive Learning [78.8942067357231]
This study investigates contrastive learning as a potential method to improve existing MPA systems.
We introduce a weighted contrastive loss suitable for regression tasks applied to a convolutional neural network.
Our results show that contrastive-based methods are able to match and exceed SoTA performance for MPA regression tasks.
arXiv Detail & Related papers (2021-08-03T19:24:25Z) - Rethinking Deep Contrastive Learning with Embedding Memory [58.66613563148031]
Pair-wise loss functions have been extensively studied and shown to continuously improve the performance of deep metric learning (DML)
We provide a new methodology for systematically studying weighting strategies of various pair-wise loss functions, and rethink pair weighting with an embedding memory.
arXiv Detail & Related papers (2021-03-25T17:39:34Z) - Learning Stereo Matchability in Disparity Regression Networks [40.08209864470944]
This paper proposes a stereo matching network that considers pixel-wise matchability.
The proposed deep stereo matchability (DSM) framework can improve the matching result or accelerate the computation while still guaranteeing the quality.
arXiv Detail & Related papers (2020-08-11T15:55:49Z) - Deep Semantic Matching with Foreground Detection and Cycle-Consistency [103.22976097225457]
We address weakly supervised semantic matching based on a deep network.
We explicitly estimate the foreground regions to suppress the effect of background clutter.
We develop cycle-consistent losses to enforce the predicted transformations across multiple images to be geometrically plausible and consistent.
arXiv Detail & Related papers (2020-03-31T22:38:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.