BCLNet: Bilateral Consensus Learning for Two-View Correspondence Pruning
- URL: http://arxiv.org/abs/2401.03459v1
- Date: Sun, 7 Jan 2024 11:38:15 GMT
- Title: BCLNet: Bilateral Consensus Learning for Two-View Correspondence Pruning
- Authors: Xiangyang Miao, Guobao Xiao, Shiping Wang, Jun Yu
- Abstract summary: Correspondence pruning aims to establish reliable correspondences between two related images.
Existing approaches often employ a progressive strategy to handle the local and global contexts.
We propose a parallel context learning strategy that involves acquiring bilateral consensus for the two-view correspondence pruning task.
- Score: 26.400567961735234
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Correspondence pruning aims to establish reliable correspondences between two
related images and recover relative camera motion. Existing approaches often
employ a progressive strategy to handle the local and global contexts, with a
prominent emphasis on transitioning from local to global, resulting in the
neglect of interactions between different contexts. To tackle this issue, we
propose a parallel context learning strategy that involves acquiring bilateral
consensus for the two-view correspondence pruning task. In our approach, we
design a distinctive self-attention block to capture global context and
parallel process it with the established local context learning module, which
enables us to simultaneously capture both local and global consensuses. By
combining these local and global consensuses, we derive the required bilateral
consensus. We also design a recalibration block, reducing the influence of
erroneous consensus information and enhancing the robustness of the model. The
culmination of our efforts is the Bilateral Consensus Learning Network
(BCLNet), which efficiently estimates camera pose and identifies inliers (true
correspondences). Extensive experiments results demonstrate that our network
not only surpasses state-of-the-art methods on benchmark datasets but also
showcases robust generalization abilities across various feature extraction
techniques. Noteworthily, BCLNet obtains 3.98\% mAP5$^{\circ}$ gains over the
second best method on unknown outdoor dataset, and obviously accelerates model
training speed. The source code will be available at:
https://github.com/guobaoxiao/BCLNet.
Related papers
- Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text Matching [53.05954114863596]
We propose a brand-new Deep Boosting Learning (DBL) algorithm for image-text matching.
An anchor branch is first trained to provide insights into the data properties.
A target branch is concurrently tasked with more adaptive margin constraints to further enlarge the relative distance between matched and unmatched samples.
arXiv Detail & Related papers (2024-04-28T08:44:28Z) - Mutual Guidance and Residual Integration for Image Enhancement [43.282397174228116]
We propose a novel mutual guidance network (MGN) to perform effective bidirectional global-local information exchange.
In our design, we adopt a two-branch framework where one branch focuses more on modeling global relations while the other is committed to processing local information.
As a result, both the global and local branches can enjoy the merits of mutual information aggregation.
arXiv Detail & Related papers (2022-11-25T06:12:39Z) - PointCMC: Cross-Modal Multi-Scale Correspondences Learning for Point
Cloud Understanding [0.875967561330372]
Cross-modal method to model multi-scale correspondences across modalities for self-supervised point cloud representation learning.
PointCMC is composed of: (1) a local-to-local (L2L) module that learns local correspondences through optimized cross-modal local geometric features, (2) a local-to-global (L2G) module that aims to learn the correspondences between local and global features across modalities via local-global discrimination, and (3) a global-to-global (G2G) module, which leverages auxiliary global contrastive loss between the point cloud and image to learn high-level semantic correspondences.
arXiv Detail & Related papers (2022-11-22T06:08:43Z) - Adaptive Local-Component-aware Graph Convolutional Network for One-shot
Skeleton-based Action Recognition [54.23513799338309]
We present an Adaptive Local-Component-aware Graph Convolutional Network for skeleton-based action recognition.
Our method provides a stronger representation than the global embedding and helps our model reach state-of-the-art.
arXiv Detail & Related papers (2022-09-21T02:33:07Z) - Global-and-Local Collaborative Learning for Co-Salient Object Detection [162.62642867056385]
The goal of co-salient object detection (CoSOD) is to discover salient objects that commonly appear in a query group containing two or more relevant images.
We propose a global-and-local collaborative learning architecture, which includes a global correspondence modeling (GCM) and a local correspondence modeling (LCM)
The proposed GLNet is evaluated on three prevailing CoSOD benchmark datasets, demonstrating that our model trained on a small dataset (about 3k images) still outperforms eleven state-of-the-art competitors trained on some large datasets (about 8k-200k images)
arXiv Detail & Related papers (2022-04-19T14:32:41Z) - DenseGAP: Graph-Structured Dense Correspondence Learning with Anchor
Points [15.953570826460869]
Establishing dense correspondence between two images is a fundamental computer vision problem.
We introduce DenseGAP, a new solution for efficient Dense correspondence learning with a Graph-structured neural network conditioned on Anchor Points.
Our method advances the state-of-the-art of correspondence learning on most benchmarks.
arXiv Detail & Related papers (2021-12-13T18:59:30Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Enriching Local and Global Contexts for Temporal Action Localization [0.0]
We enrich both the local and global contexts in the popular two-stage temporal localization framework.
Our proposed model, dubbed ContextLoc, can be divided into three sub-networks: L-Net, G-Net and P-Net.
The efficacy of our proposed method is validated by experimental results on the THUMOS14 (54.3% at IoU@0.5) and ActivityNet v1.3 (51.24% at IoU@0.5) datasets.
arXiv Detail & Related papers (2021-07-27T17:25:40Z) - Global Context Aware RCNN for Object Detection [1.1939762265857436]
We propose a novel end-to-end trainable framework, called Global Context Aware (GCA) RCNN.
The core component of GCA framework is a context aware mechanism, in which both global feature pyramid and attention strategies are used for feature extraction and feature refinement.
In the end, we also present a lightweight version of our method, which only slightly increases model complexity and computational burden.
arXiv Detail & Related papers (2020-12-04T14:56:46Z) - Disentangled Non-Local Neural Networks [68.92293183542131]
We study the non-local block in depth, where we find that its attention can be split into two terms.
We present the disentangled non-local block, where the two terms are decoupled to facilitate learning for both terms.
arXiv Detail & Related papers (2020-06-11T17:59:22Z) - Cascaded Human-Object Interaction Recognition [175.60439054047043]
We introduce a cascade architecture for a multi-stage, coarse-to-fine HOI understanding.
At each stage, an instance localization network progressively refines HOI proposals and feeds them into an interaction recognition network.
With our carefully-designed human-centric relation features, these two modules work collaboratively towards effective interaction understanding.
arXiv Detail & Related papers (2020-03-09T17:05:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.