PairingNet: A Learning-based Pair-searching and -matching Network for
Image Fragments
- URL: http://arxiv.org/abs/2312.08704v1
- Date: Thu, 14 Dec 2023 07:43:53 GMT
- Title: PairingNet: A Learning-based Pair-searching and -matching Network for
Image Fragments
- Authors: Rixin Zhou, Ding Xia, Yi Zhang, Honglin Pang, Xi Yang, Chuntao Li
- Abstract summary: We propose a learning-based image fragment pair-searching and -matching approach to solve the challenging restoration problem.
Our proposed network achieves excellent pair-searching accuracy, reduces matching errors, and significantly reduces computational time.
- Score: 6.694162736590122
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we propose a learning-based image fragment pair-searching and
-matching approach to solve the challenging restoration problem. Existing works
use rule-based methods to match similar contour shapes or textures, which are
always difficult to tune hyperparameters for extensive data and computationally
time-consuming. Therefore, we propose a neural network that can effectively
utilize neighbor textures with contour shape information to fundamentally
improve performance. First, we employ a graph-based network to extract the
local contour and texture features of fragments. Then, for the pair-searching
task, we adopt a linear transformer-based module to integrate these local
features and use contrastive loss to encode the global features of each
fragment. For the pair-matching task, we design a weighted fusion module to
dynamically fuse extracted local contour and texture features, and formulate a
similarity matrix for each pair of fragments to calculate the matching score
and infer the adjacent segment of contours. To faithfully evaluate our proposed
network, we created a new image fragment dataset through an algorithm we
designed that tears complete images into irregular fragments. The experimental
results show that our proposed network achieves excellent pair-searching
accuracy, reduces matching errors, and significantly reduces computational
time. Details, sourcecode, and data are available in our supplementary
material.
Related papers
- FuseNet: Self-Supervised Dual-Path Network for Medical Image
Segmentation [3.485615723221064]
FuseNet is a dual-stream framework for self-supervised semantic segmentation.
Cross-modal fusion technique extends the principles of CLIP by replacing textual data with augmented images.
experiments on skin lesion and lung segmentation datasets demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2023-11-22T00:03:16Z) - Efficient Segmentation with Texture in Ore Images Based on
Box-supervised Approach [6.6773975364173]
A box-supervised technique with texture features is proposed to identify complete and independent ores.
The proposed method achieves over 50 frames per second with a small model size of 21.6 MB.
The method maintains a high level of accuracy compared with the state-of-the-art approaches on ore image dataset.
arXiv Detail & Related papers (2023-11-10T08:28:22Z) - High-resolution Iterative Feedback Network for Camouflaged Object
Detection [128.893782016078]
Spotting camouflaged objects that are visually assimilated into the background is tricky for object detection algorithms.
We aim to extract the high-resolution texture details to avoid the detail degradation that causes blurred vision in edges and boundaries.
We introduce a novel HitNet to refine the low-resolution representations by high-resolution features in an iterative feedback manner.
arXiv Detail & Related papers (2022-03-22T11:20:21Z) - Compositional Sketch Search [91.84489055347585]
We present an algorithm for searching image collections using free-hand sketches.
We exploit drawings as a concise and intuitive representation for specifying entire scene compositions.
arXiv Detail & Related papers (2021-06-15T09:38:09Z) - DFM: A Performance Baseline for Deep Feature Matching [10.014010310188821]
The proposed method uses pre-trained VGG architecture as a feature extractor and does not require any additional training specific to improve matching.
Our algorithm achieves 0.57 and 0.80 overall scores in terms of Mean Matching Accuracy (MMA) for 1 pixel and 2 pixels thresholds respectively on Hpatches dataset.
arXiv Detail & Related papers (2021-06-14T22:55:06Z) - Image Retrieval for Structure-from-Motion via Graph Convolutional
Network [13.040952255039702]
We present a novel retrieval method based on Graph Convolutional Network (GCN) to generate accurate pairwise matches without costly redundancy.
By constructing a subgraph surrounding the query image as input data, we adopt a learnable GCN to exploit whether nodes in the subgraph have overlapping regions with the query photograph.
Experiments demonstrate that our method performs remarkably well on the challenging dataset of highly ambiguous and duplicated scenes.
arXiv Detail & Related papers (2020-09-17T04:03:51Z) - GOCor: Bringing Globally Optimized Correspondence Volumes into Your
Neural Network [176.3781969089004]
Feature correlation layer serves as a key neural network module in computer vision problems that involve dense correspondences between image pairs.
We propose GOCor, a fully differentiable dense matching module, acting as a direct replacement to the feature correlation layer.
Our approach significantly outperforms the feature correlation layer for the tasks of geometric matching, optical flow, and dense semantic matching.
arXiv Detail & Related papers (2020-09-16T17:33:01Z) - Neural Subdivision [58.97214948753937]
This paper introduces Neural Subdivision, a novel framework for data-driven coarseto-fine geometry modeling.
We optimize for the same set of network weights across all local mesh patches, thus providing an architecture that is not constrained to a specific input mesh, fixed genus, or category.
We demonstrate that even when trained on a single high-resolution mesh our method generates reasonable subdivisions for novel shapes.
arXiv Detail & Related papers (2020-05-04T20:03:21Z) - Adaptive Fractional Dilated Convolution Network for Image Aesthetics
Assessment [33.945579916184364]
An adaptive fractional dilated convolution (AFDC) is developed to tackle this issue in convolutional kernel level.
We provide a concise formulation for mini-batch training and utilize a grouping strategy to reduce computational overhead.
Our experimental results demonstrate that our proposed method achieves state-of-the-art performance on image aesthetics assessment over the AVA dataset.
arXiv Detail & Related papers (2020-04-06T21:56:29Z) - Weakly-Supervised Semantic Segmentation by Iterative Affinity Learning [86.45526827323954]
Weakly-supervised semantic segmentation is a challenging task as no pixel-wise label information is provided for training.
We propose an iterative algorithm to learn such pairwise relations.
We show that the proposed algorithm performs favorably against the state-of-the-art methods.
arXiv Detail & Related papers (2020-02-19T10:32:03Z) - Image Fine-grained Inpainting [89.17316318927621]
We present a one-stage model that utilizes dense combinations of dilated convolutions to obtain larger and more effective receptive fields.
To better train this efficient generator, except for frequently-used VGG feature matching loss, we design a novel self-guided regression loss.
We also employ a discriminator with local and global branches to ensure local-global contents consistency.
arXiv Detail & Related papers (2020-02-07T03:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.