R2FD2: Fast and Robust Matching of Multimodal Remote Sensing Image via
Repeatable Feature Detector and Rotation-invariant Feature Descriptor
- URL: http://arxiv.org/abs/2212.02277v2
- Date: Tue, 6 Dec 2022 07:05:04 GMT
- Title: R2FD2: Fast and Robust Matching of Multimodal Remote Sensing Image via
Repeatable Feature Detector and Rotation-invariant Feature Descriptor
- Authors: Bai Zhu, Chao Yang, Jinkun Dai, Jianwei Fan, Yuanxin Ye
- Abstract summary: We propose a novel feature matching method (named R2FD2) that is robust to radiation and rotation differences.
The proposed R2FD2 outperforms five state-of-the-art feature matching methods, and has superior advantages in universality and adaptability.
Our R2FD2 achieves the accuracy of matching within two pixels and has a great advantage in matching efficiency over other state-of-the-art methods.
- Score: 3.395266574804949
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automatically identifying feature correspondences between multimodal images
is facing enormous challenges because of the significant differences both in
radiation and geometry. To address these problems, we propose a novel feature
matching method (named R2FD2) that is robust to radiation and rotation
differences. Our R2FD2 is conducted in two critical contributions, consisting
of a repeatable feature detector and a rotation-invariant feature descriptor.
In the first stage, a repeatable feature detector called the Multi-channel
Auto-correlation of the Log-Gabor (MALG) is presented for feature detection,
which combines the multi-channel auto-correlation strategy with the Log-Gabor
wavelets to detect interest points (IPs) with high repeatability and uniform
distribution. In the second stage, a rotation-invariant feature descriptor is
constructed, named the Rotation-invariant Maximum index map of the Log-Gabor
(RMLG), which consists of two components: fast assignment of dominant
orientation and construction of feature representation. In the process of fast
assignment of dominant orientation, a Rotation-invariant Maximum Index Map
(RMIM) is built to address rotation deformations. Then, the proposed RMLG
incorporates the rotation-invariant RMIM with the spatial configuration of
DAISY to depict a more discriminative feature representation, which improves
RMLG's resistance to radiation and rotation variances.Experimental results show
that the proposed R2FD2 outperforms five state-of-the-art feature matching
methods, and has superior advantages in adaptability and universality.
Moreover, our R2FD2 achieves the accuracy of matching within two pixels and has
a great advantage in matching efficiency over other state-of-the-art methods.
Related papers
- Exploring Invariant Representation for Visible-Infrared Person
Re-Identification [77.06940947765406]
Cross-spectral person re-identification, which aims to associate identities to pedestrians across different spectra, faces a main challenge of the modality discrepancy.
In this paper, we address the problem from both image-level and feature-level in an end-to-end hybrid learning framework named robust feature mining network (RFM)
Experiment results on two standard cross-spectral person re-identification datasets, RegDB and SYSU-MM01, have demonstrated state-of-the-art performance.
arXiv Detail & Related papers (2023-02-02T05:24:50Z) - Detecting Rotated Objects as Gaussian Distributions and Its 3-D
Generalization [81.29406957201458]
Existing detection methods commonly use a parameterized bounding box (BBox) to model and detect (horizontal) objects.
We argue that such a mechanism has fundamental limitations in building an effective regression loss for rotation detection.
We propose to model the rotated objects as Gaussian distributions.
We extend our approach from 2-D to 3-D with a tailored algorithm design to handle the heading estimation.
arXiv Detail & Related papers (2022-09-22T07:50:48Z) - Look Back and Forth: Video Super-Resolution with Explicit Temporal
Difference Modeling [105.69197687940505]
We propose to explore the role of explicit temporal difference modeling in both LR and HR space.
To further enhance the super-resolution result, not only spatial residual features are extracted, but the difference between consecutive frames in high-frequency domain is also computed.
arXiv Detail & Related papers (2022-04-14T17:07:33Z) - A Robust Multimodal Remote Sensing Image Registration Method and System
Using Steerable Filters with First- and Second-order Gradients [7.813406811407584]
Co-registration of multimodal remote sensing images is still an ongoing challenge because of nonlinear radiometric differences (NRD) and significant geometric distortions.
In this paper, a robust matching method based on the Steerable filters is proposed consisting of two critical steps.
The performance of the proposed matching method has been evaluated with many different kinds of multimodal RS images.
arXiv Detail & Related papers (2022-02-27T12:22:42Z) - Cross-Modality Attentive Feature Fusion for Object Detection in
Multispectral Remote Sensing Imagery [0.6853165736531939]
Cross-modality fusing complementary information of multispectral remote sensing image pairs can improve the perception ability of detection algorithms.
We propose a novel and lightweight multispectral feature fusion approach with joint common-modality and differential-modality attentions.
Our proposed approach can achieve the state-of-the-art performance at a low cost.
arXiv Detail & Related papers (2021-12-06T13:12:36Z) - BAANet: Learning Bi-directional Adaptive Attention Gates for
Multispectral Pedestrian Detection [14.672188805059744]
This work proposes an effective and efficient cross-modality fusion module called Bi-directional Adaptive Gate (BAA-Gate)
Based on the attention mechanism, the BAA-Gate is devised to distill the informative features and recalibrate the representationsally.
Considerable experiments on the challenging KAIST dataset demonstrate the superior performance of our method with satisfactory speed.
arXiv Detail & Related papers (2021-12-04T08:30:54Z) - Transformer-based Network for RGB-D Saliency Detection [82.6665619584628]
Key to RGB-D saliency detection is to fully mine and fuse information at multiple scales across the two modalities.
We show that transformer is a uniform operation which presents great efficacy in both feature fusion and feature enhancement.
Our proposed network performs favorably against state-of-the-art RGB-D saliency detection methods.
arXiv Detail & Related papers (2021-12-01T15:53:58Z) - Robust Reference-based Super-Resolution via C2-Matching [77.51610726936657]
Super-Resolution (Ref-SR) has recently emerged as a promising paradigm to enhance a low-resolution (LR) input image by introducing an additional high-resolution (HR) reference image.
Existing Ref-SR methods mostly rely on implicit correspondence matching to borrow HR textures from reference images to compensate for the information loss in input images.
We propose C2-Matching, which produces explicit robust matching crossing transformation and resolution.
arXiv Detail & Related papers (2021-06-03T16:40:36Z) - AFD-Net: Adaptive Fully-Dual Network for Few-Shot Object Detection [8.39479809973967]
Few-shot object detection (FSOD) aims at learning a detector that can fast adapt to previously unseen objects with scarce examples.
Existing methods solve this problem by performing subtasks of classification and localization utilizing a shared component.
We present that a general few-shot detector should consider the explicit decomposition of two subtasks, as well as leveraging information from both of them to enhance feature representations.
arXiv Detail & Related papers (2020-11-30T10:21:32Z) - Directional ASR: A New Paradigm for E2E Multi-Speaker Speech Recognition
with Source Localization [73.62550438861942]
This paper proposes a new paradigm for handling far-field multi-speaker data in an end-to-end neural network manner, called directional automatic speech recognition (D-ASR)
In D-ASR, the azimuth angle of the sources with respect to the microphone array is defined as a latent variable. This angle controls the quality of separation, which in turn determines the ASR performance.
arXiv Detail & Related papers (2020-10-30T20:26:28Z) - A Dual Iterative Refinement Method for Non-rigid Shape Matching [16.03666555216332]
A simple and efficient dual iterative refinement (DIR) method is proposed for dense correspondence between two nearly isometric shapes.
The key idea is to use dual information, such as spatial and spectral, or local and global features, in a complementary and effective way.
Experiments on various data sets demonstrate the superiority of DIR over other state-of-the-art methods in terms of both accuracy and efficiency.
arXiv Detail & Related papers (2020-07-26T03:35:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.