Local Similarity Pattern and Cost Self-Reassembling for Deep Stereo
Matching Networks
- URL: http://arxiv.org/abs/2112.01011v2
- Date: Sun, 5 Dec 2021 01:38:31 GMT
- Title: Local Similarity Pattern and Cost Self-Reassembling for Deep Stereo
Matching Networks
- Authors: Biyang Liu, Huimin Yu, Yangqi Long
- Abstract summary: We introduce a pairwise feature for deep stereo matching networks, named LSP (Local Similarity Pattern)
Through explicitly revealing the neighbor relationships, LSP contains rich structural information, which can be leveraged to aid for more discriminative feature description.
Secondly, we design a dynamic self-reassembling refinement strategy and apply it to the cost distribution and the disparity map respectively.
- Score: 3.7384509727711923
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although convolution neural network based stereo matching architectures have
made impressive achievements, there are still some limitations: 1)
Convolutional Feature (CF) tends to capture appearance information, which is
inadequate for accurate matching. 2) Due to the static filters, current
convolution based disparity refinement modules often produce over-smooth
results. In this paper, we present two schemes to address these issues, where
some traditional wisdoms are integrated. Firstly, we introduce a pairwise
feature for deep stereo matching networks, named LSP (Local Similarity
Pattern). Through explicitly revealing the neighbor relationships, LSP contains
rich structural information, which can be leveraged to aid CF for more
discriminative feature description. Secondly, we design a dynamic
self-reassembling refinement strategy and apply it to the cost distribution and
the disparity map respectively. The former could be equipped with the unimodal
distribution constraint to alleviate the over-smoothing problem, and the latter
is more practical. The effectiveness of the proposed methods is demonstrated
via incorporating them into two well-known basic architectures, GwcNet and
GANet-deep. Experimental results on the SceneFlow and KITTI benchmarks show
that our modules significantly improve the performance of the model.
Related papers
- A Noise and Edge extraction-based dual-branch method for Shallowfake and Deepfake Localization [15.647035299476894]
We develop a dual-branch model that integrates manually designed feature noise with conventional CNN features.
The model is superior in comparison and easily outperforms the existing state-of-the-art (SoTA) models.
arXiv Detail & Related papers (2024-09-02T02:18:34Z) - Double-Shot 3D Shape Measurement with a Dual-Branch Network [14.749887303860717]
We propose a dual-branch Convolutional Neural Network (CNN)-Transformer network (PDCNet) to process different structured light (SL) modalities.
Within PDCNet, a Transformer branch is used to capture global perception in the fringe images, while a CNN branch is designed to collect local details in the speckle images.
We show that our method can reduce fringe order ambiguity while producing high-accuracy results on a self-made dataset.
arXiv Detail & Related papers (2024-07-19T10:49:26Z) - A Refreshed Similarity-based Upsampler for Direct High-Ratio Feature Upsampling [54.05517338122698]
We propose an explicitly controllable query-key feature alignment from both semantic-aware and detail-aware perspectives.
We also develop a fine-grained neighbor selection strategy on HR features, which is simple yet effective for alleviating mosaic artifacts.
Our proposed ReSFU framework consistently achieves satisfactory performance on different segmentation applications.
arXiv Detail & Related papers (2024-07-02T14:12:21Z) - Trading-off Mutual Information on Feature Aggregation for Face
Recognition [12.803514943105657]
We propose a technique to aggregate the outputs of two state-of-the-art (SOTA) deep Face Recognition (FR) models.
In our approach, we leverage the transformer attention mechanism to exploit the relationship between different parts of two feature maps.
To evaluate the effectiveness of our proposed method, we conducted experiments on popular benchmarks and compared our results with state-of-the-art algorithms.
arXiv Detail & Related papers (2023-09-22T18:48:38Z) - Adaptive Spot-Guided Transformer for Consistent Local Feature Matching [64.30749838423922]
We propose Adaptive Spot-Guided Transformer (ASTR) for local feature matching.
ASTR models the local consistency and scale variations in a unified coarse-to-fine architecture.
arXiv Detail & Related papers (2023-03-29T12:28:01Z) - Semantic Feature Integration network for Fine-grained Visual
Classification [5.182627302449368]
We propose the Semantic Feature Integration network (SFI-Net) to address the above difficulties.
By eliminating unnecessary features and reconstructing the semantic relations among discriminative features, our SFI-Net has achieved satisfying performance.
arXiv Detail & Related papers (2023-02-13T07:32:25Z) - BCFNet: A Balanced Collaborative Filtering Network with Attention
Mechanism [106.43103176833371]
Collaborative Filtering (CF) based recommendation methods have been widely studied.
We propose a novel recommendation model named Balanced Collaborative Filtering Network (BCFNet)
In addition, an attention mechanism is designed to better capture the hidden information within implicit feedback and strengthen the learning ability of the neural network.
arXiv Detail & Related papers (2021-03-10T14:59:23Z) - Learning to Generate Content-Aware Dynamic Detectors [62.74209921174237]
We introduce a newpective of designing efficient detectors, which is automatically generating sample-adaptive model architecture.
We introduce a course-to-fine strat-egy tailored for object detection to guide the learning of dynamic routing.
Experiments on MS-COCO dataset demonstrate that CADDet achieves 1.8 higher mAP with 10% fewer FLOPs compared with vanilla routing.
arXiv Detail & Related papers (2020-12-08T08:05:20Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z) - The Heterogeneity Hypothesis: Finding Layer-Wise Differentiated Network
Architectures [179.66117325866585]
We investigate a design space that is usually overlooked, i.e. adjusting the channel configurations of predefined networks.
We find that this adjustment can be achieved by shrinking widened baseline networks and leads to superior performance.
Experiments are conducted on various networks and datasets for image classification, visual tracking and image restoration.
arXiv Detail & Related papers (2020-06-29T17:59:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.