DRKF: Distilled Rotated Kernel Fusion for Efficient Rotation Invariant
Descriptors in Local Feature Matching
- URL: http://arxiv.org/abs/2209.10907v3
- Date: Fri, 5 Jan 2024 12:03:22 GMT
- Title: DRKF: Distilled Rotated Kernel Fusion for Efficient Rotation Invariant
Descriptors in Local Feature Matching
- Authors: Ranran Huang, Jiancheng Cai, Chao Li, Zhuoyuan Wu, Xinmin Liu, Zhenhua
Chai
- Abstract summary: Rotated Fusion Kernel (RKF) imposes rotations on the convolution kernel to improve the inherent nature of CNN.
MOFA aggregates features extracted from multiple rotated versions of the input image.
Our method can outperform other state-of-the-art techniques when exposed to large rotation variations.
- Score: 9.68840174997957
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The performance of local feature descriptors degrades in the presence of
large rotation variations. To address this issue, we present an efficient
approach to learning rotation invariant descriptors. Specifically, we propose
Rotated Kernel Fusion (RKF) which imposes rotations on the convolution kernel
to improve the inherent nature of CNN. Since RKF can be processed by the
subsequent re-parameterization, no extra computational costs will be introduced
in the inference stage. Moreover, we present Multi-oriented Feature Aggregation
(MOFA) which aggregates features extracted from multiple rotated versions of
the input image and can provide auxiliary knowledge for the training of RKF by
leveraging the distillation strategy. We refer to the distilled RKF model as
DRKF. Besides the evaluation on a rotation-augmented version of the public
dataset HPatches, we also contribute a new dataset named DiverseBEV which is
collected during the drone's flight and consists of bird's eye view images with
large viewpoint changes and camera rotations. Extensive experiments show that
our method can outperform other state-of-the-art techniques when exposed to
large rotation variations.
Related papers
- Exploring Kernel Transformations for Implicit Neural Representations [57.2225355625268]
Implicit neural representations (INRs) leverage neural networks to represent signals by mapping coordinates to their corresponding attributes.
This work pioneers the exploration of the effect of kernel transformation of input/output while keeping the model itself unchanged.
A byproduct of our findings is a simple yet effective method that combines scale and shift to significantly boost INR with negligible overhead.
arXiv Detail & Related papers (2025-04-07T04:43:50Z) - Variance-Reducing Couplings for Random Features [57.73648780299374]
Random features (RFs) are a popular technique to scale up kernel methods in machine learning.
We find couplings to improve RFs defined on both Euclidean and discrete input spaces.
We reach surprising conclusions about the benefits and limitations of variance reduction as a paradigm.
arXiv Detail & Related papers (2024-05-26T12:25:09Z) - FRED: Towards a Full Rotation-Equivariance in Aerial Image Object
Detection [28.47314201641291]
We introduce a Fully Rotation-Equivariant Oriented Object Detector (FRED)
Our proposed method delivers comparable performance on DOTA-v1.0 and outperforms by 1.5 mAP on DOTA-v1.5, all while significantly reducing the model parameters to 16%.
arXiv Detail & Related papers (2023-12-22T09:31:43Z) - Sorted Convolutional Network for Achieving Continuous Rotational
Invariance [56.42518353373004]
We propose a Sorting Convolution (SC) inspired by some hand-crafted features of texture images.
SC achieves continuous rotational invariance without requiring additional learnable parameters or data augmentation.
Our results demonstrate that SC achieves the best performance in the aforementioned tasks.
arXiv Detail & Related papers (2023-05-23T18:37:07Z) - CRIN: Rotation-Invariant Point Cloud Analysis and Rotation Estimation
via Centrifugal Reference Frame [60.24797081117877]
We propose the CRIN, namely Centrifugal Rotation-Invariant Network.
CRIN directly takes the coordinates of points as input and transforms local points into rotation-invariant representations.
A continuous distribution for 3D rotations based on points is introduced.
arXiv Detail & Related papers (2023-03-06T13:14:10Z) - RIC-CNN: Rotation-Invariant Coordinate Convolutional Neural Network [56.42518353373004]
We propose a new convolutional operation, called Rotation-Invariant Coordinate Convolution (RIC-C)
By replacing all standard convolutional layers in a CNN with the corresponding RIC-C, a RIC-CNN can be derived.
It can be observed that RIC-CNN achieves the state-of-the-art classification on the rotated test dataset of MNIST.
arXiv Detail & Related papers (2022-11-21T19:27:02Z) - ReAFFPN: Rotation-equivariant Attention Feature Fusion Pyramid Networks
for Aerial Object Detection [13.12300373217556]
This paper proposes a Rotation-equivariant Attention Feature Fusion Pyramid Networks for Aerial Object Detection named ReAFFPN.
ReAFFPN aims at improving the effect of rotation-equivariant features fusion between adjacent layers which suffers from the semantic and scale discontinuity.
Experimental results demonstrate that ReAFFPN achieves a better rotation-equivariant feature fusion ability and significantly improve the accuracy of the Rotation-equivariant Convolutional Networks.
arXiv Detail & Related papers (2022-10-17T03:11:45Z) - Orthonormal Convolutions for the Rotation Based Iterative
Gaussianization [64.44661342486434]
This paper elaborates an extension of rotation-based iterative Gaussianization, RBIG, which makes image Gaussianization possible.
In images its application has been restricted to small image patches or isolated pixels, because rotation in RBIG is based on principal or independent component analysis.
We present the emphConvolutional RBIG: an extension that alleviates this issue by imposing that the rotation in RBIG is a convolution.
arXiv Detail & Related papers (2022-06-08T12:56:34Z) - ReF -- Rotation Equivariant Features for Local Feature Matching [30.459559206664427]
We propose an alternative, complementary approach that centers on inducing bias in the model architecture itself to generate rotation-specific' features.
We demonstrate that this high performance, rotation-specific coverage from the steerable CNNs can be expanded to all rotation angles.
We present a detailed analysis of the performance effects of ensembling, robust estimation, network architecture variations, and the use of rotation priors.
arXiv Detail & Related papers (2022-03-10T07:36:09Z) - ART-Point: Improving Rotation Robustness of Point Cloud Classifiers via
Adversarial Rotation [89.47574181669903]
In this study, we show that the rotation robustness of point cloud classifiers can also be acquired via adversarial training.
Specifically, our proposed framework named ART-Point regards the rotation of the point cloud as an attack.
We propose a fast one-step optimization to efficiently reach the final robust model.
arXiv Detail & Related papers (2022-03-08T07:20:16Z) - Attentive Rotation Invariant Convolution for Point Cloud-based Large
Scale Place Recognition [11.433270318356675]
We propose an Attentive Rotation Invariant Convolution (ARIConv) in this paper.
We experimentally demonstrate that our model can achieve state-of-the-art performance on large scale place recognition task when the point cloud scans are rotated.
arXiv Detail & Related papers (2021-08-29T09:10:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.