ReAFFPN: Rotation-equivariant Attention Feature Fusion Pyramid Networks
for Aerial Object Detection
- URL: http://arxiv.org/abs/2210.08715v1
- Date: Mon, 17 Oct 2022 03:11:45 GMT
- Title: ReAFFPN: Rotation-equivariant Attention Feature Fusion Pyramid Networks
for Aerial Object Detection
- Authors: Chongyu Sun, Yang Xu, Zebin Wu, Zhihui Wei
- Abstract summary: This paper proposes a Rotation-equivariant Attention Feature Fusion Pyramid Networks for Aerial Object Detection named ReAFFPN.
ReAFFPN aims at improving the effect of rotation-equivariant features fusion between adjacent layers which suffers from the semantic and scale discontinuity.
Experimental results demonstrate that ReAFFPN achieves a better rotation-equivariant feature fusion ability and significantly improve the accuracy of the Rotation-equivariant Convolutional Networks.
- Score: 13.12300373217556
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper proposes a Rotation-equivariant Attention Feature Fusion Pyramid
Networks for Aerial Object Detection named ReAFFPN. ReAFFPN aims at improving
the effect of rotation-equivariant features fusion between adjacent layers
which suffers from the semantic and scale discontinuity. Due to the
particularity of rotational equivariant convolution, general methods are unable
to achieve their original effect while ensuring rotation equivariance of the
network. To solve this problem, we design a new Rotation-equivariant Channel
Attention which has the ability to both generate channel attention and keep
rotation equivariance. Then we embed a new channel attention function into
Iterative Attentional Feature Fusion (iAFF) module to realize
Rotation-equivariant Attention Feature Fusion. Experimental results demonstrate
that ReAFFPN achieves a better rotation-equivariant feature fusion ability and
significantly improve the accuracy of the Rotation-equivariant Convolutional
Networks.
Related papers
- Exploring Kernel Transformations for Implicit Neural Representations [57.2225355625268]
Implicit neural representations (INRs) leverage neural networks to represent signals by mapping coordinates to their corresponding attributes.
This work pioneers the exploration of the effect of kernel transformation of input/output while keeping the model itself unchanged.
A byproduct of our findings is a simple yet effective method that combines scale and shift to significantly boost INR with negligible overhead.
arXiv Detail & Related papers (2025-04-07T04:43:50Z) - PARE-Net: Position-Aware Rotation-Equivariant Networks for Robust Point Cloud Registration [8.668461141536383]
Learning rotation-invariant distinctive features is a fundamental requirement for point cloud registration.
Existing methods often use rotation-sensitive networks to extract features, while employing rotation augmentation to learn an approximate invariant mapping rudely.
We propose a novel position-aware rotation-equivariant network, for efficient, light-weighted, and robust registration.
arXiv Detail & Related papers (2024-07-14T10:26:38Z) - FRED: Towards a Full Rotation-Equivariance in Aerial Image Object
Detection [28.47314201641291]
We introduce a Fully Rotation-Equivariant Oriented Object Detector (FRED)
Our proposed method delivers comparable performance on DOTA-v1.0 and outperforms by 1.5 mAP on DOTA-v1.5, all while significantly reducing the model parameters to 16%.
arXiv Detail & Related papers (2023-12-22T09:31:43Z) - Convolution and Attention Mixer for Synthetic Aperture Radar Image
Change Detection [41.38587746899477]
Synthetic aperture radar (SAR) image change detection is a critical task and has received increasing attentions in the remote sensing community.
Existing SAR change detection methods are mainly based on convolutional neural networks (CNNs)
We propose a convolution and attention mixer (CAMixer) to incorporate global attention.
arXiv Detail & Related papers (2023-09-21T12:28:23Z) - VI-Net: Boosting Category-level 6D Object Pose Estimation via Learning
Decoupled Rotations on the Spherical Representations [55.25238503204253]
We propose a novel rotation estimation network, termed as VI-Net, to make the task easier.
To process the spherical signals, a Spherical Feature Pyramid Network is constructed based on a novel design of SPAtial Spherical Convolution.
Experiments on the benchmarking datasets confirm the efficacy of our method, which outperforms the existing ones with a large margin in the regime of high precision.
arXiv Detail & Related papers (2023-08-19T05:47:53Z) - Rotation-Invariant Transformer for Point Cloud Matching [42.5714375149213]
We introduce RoITr, a Rotation-Invariant Transformer to cope with the pose variations in the point cloud matching task.
We propose a global transformer with rotation-invariant cross-frame spatial awareness learned by the self-attention mechanism.
RoITr surpasses the existing methods by at least 13 and 5 percentage points in terms of Inlier Ratio and Registration Recall.
arXiv Detail & Related papers (2023-03-14T20:55:27Z) - CRIN: Rotation-Invariant Point Cloud Analysis and Rotation Estimation
via Centrifugal Reference Frame [60.24797081117877]
We propose the CRIN, namely Centrifugal Rotation-Invariant Network.
CRIN directly takes the coordinates of points as input and transforms local points into rotation-invariant representations.
A continuous distribution for 3D rotations based on points is introduced.
arXiv Detail & Related papers (2023-03-06T13:14:10Z) - Deep Neural Networks with Efficient Guaranteed Invariances [77.99182201815763]
We address the problem of improving the performance and in particular the sample complexity of deep neural networks.
Group-equivariant convolutions are a popular approach to obtain equivariant representations.
We propose a multi-stream architecture, where each stream is invariant to a different transformation.
arXiv Detail & Related papers (2023-03-02T20:44:45Z) - DRKF: Distilled Rotated Kernel Fusion for Efficient Rotation Invariant
Descriptors in Local Feature Matching [9.68840174997957]
Rotated Fusion Kernel (RKF) imposes rotations on the convolution kernel to improve the inherent nature of CNN.
MOFA aggregates features extracted from multiple rotated versions of the input image.
Our method can outperform other state-of-the-art techniques when exposed to large rotation variations.
arXiv Detail & Related papers (2022-09-22T10:29:17Z) - RSDet++: Point-based Modulated Loss for More Accurate Rotated Object
Detection [53.57176614020894]
We classify the discontinuity of loss in both five-param and eight-param rotated object detection methods as rotation sensitivity error (RSE)
We introduce a novel modulated rotation loss to alleviate the problem and propose a rotation sensitivity detection network (RSDet)
To further improve the accuracy of our method on objects smaller than 10 pixels, we introduce a novel RSDet++.
arXiv Detail & Related papers (2021-09-24T11:57:53Z) - Deep Positional and Relational Feature Learning for Rotation-Invariant
Point Cloud Analysis [107.9979381402172]
We propose a rotation-invariant deep network for point clouds analysis.
The network is hierarchical and relies on two modules: a positional feature embedding block and a relational feature embedding block.
Experiments show state-of-the-art classification and segmentation performances on benchmark datasets.
arXiv Detail & Related papers (2020-11-18T04:16:51Z) - A Rotation-Invariant Framework for Deep Point Cloud Analysis [132.91915346157018]
We introduce a new low-level purely rotation-invariant representation to replace common 3D Cartesian coordinates as the network inputs.
Also, we present a network architecture to embed these representations into features, encoding local relations between points and their neighbors, and the global shape structure.
We evaluate our method on multiple point cloud analysis tasks, including shape classification, part segmentation, and shape retrieval.
arXiv Detail & Related papers (2020-03-16T14:04:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.