ssFPN: Scale Sequence (S^2) Feature Based-Feature Pyramid Network for
Object Detection
- URL: http://arxiv.org/abs/2208.11533v2
- Date: Thu, 25 Aug 2022 04:22:56 GMT
- Title: ssFPN: Scale Sequence (S^2) Feature Based-Feature Pyramid Network for
Object Detection
- Authors: Hye-Jin Park, Young-Ju Choi, Young-Woon Lee, Byung-Gyu Kim
- Abstract summary: We propose a new scale sequence (S2) feature extraction of Feature Pyramid Network (FPN) to strengthen feature information of small objects.
We demonstrate the proposed S2 feature can improve the performance of both one-stage and two-stage detectors on MS COCO dataset.
- Score: 4.844193288417161
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Feature Pyramid Network (FPN) has been an essential module for object
detection models to consider various scales of an object. However, average
precision (AP) on small objects is relatively lower than AP on medium and large
objects. The reason is why the deeper layer of CNN causes information loss as
feature extraction level. We propose a new scale sequence (S^2) feature
extraction of FPN to strengthen feature information of small objects. We
consider FPN structure as scale-space and extract scale sequence (S^2) feature
by 3D convolution on the level axis of FPN. It is basically scale invariant
feature and is built on high-resolution pyramid feature map for small objects.
Furthermore, the proposed S^2 feature can be extended to most object detection
models based on FPN. We demonstrate the proposed S2 feature can improve the
performance of both one-stage and two-stage detectors on MS COCO dataset. Based
on the proposed S2 feature, we achieve upto 1.3% and 1.1% of AP improvement for
YOLOv4-P5 and YOLOv4-P6, respectively. For Faster RCNN and Mask R-CNN, we
observe upto 2.0% and 1.6% of AP improvement with the suggested S^2 feature,
respectively.
Related papers
- CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds [55.44204039410225]
We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D.
Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels.
To recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module.
arXiv Detail & Related papers (2022-10-09T13:38:48Z) - SFPN: Synthetic FPN for Object Detection [6.117917355232904]
This paper proposes a new SFPN (Synthetic Fusion Pyramid Network) arichtecture to enhance the accuracy of light-weight CNN backones.
Experiments prove the SFPN architecture outperforms either the large backbone VGG16, ResNet50 or light-weight backbones such as MobilenetV2 based on AP score.
arXiv Detail & Related papers (2022-03-04T17:19:50Z) - Pyramid R-CNN: Towards Better Performance and Adaptability for 3D Object
Detection [89.66162518035144]
We present a flexible and high-performance framework, named Pyramid R-CNN, for two-stage 3D object detection from point clouds.
We propose a novel second-stage module, named pyramid RoI head, to adaptively learn the features from the sparse points of interest.
Our pyramid RoI head is robust to the sparse and imbalanced circumstances, and can be applied upon various 3D backbones to consistently boost the detection performance.
arXiv Detail & Related papers (2021-09-06T14:17:51Z) - A^2-FPN: Attention Aggregation based Feature Pyramid Network for
Instance Segmentation [68.10621089649486]
We propose Attention Aggregation based Feature Pyramid Network (A2-FPN) to improve multi-scale feature learning.
A2-FPN achieves an improvement of 2.0% and 1.4% mask AP when integrated into the strong baselines such as Cascade Mask R-CNN and Hybrid Task Cascade.
arXiv Detail & Related papers (2021-05-07T11:51:08Z) - FS-Net: Fast Shape-based Network for Category-Level 6D Object Pose
Estimation with Decoupled Rotation Mechanism [49.89268018642999]
We propose a fast shape-based network (FS-Net) with efficient category-level feature extraction for 6D pose estimation.
The proposed method achieves state-of-the-art performance in both category- and instance-level 6D object pose estimation.
arXiv Detail & Related papers (2021-03-12T03:07:24Z) - OPANAS: One-Shot Path Aggregation Network Architecture Search for Object
Detection [82.04372532783931]
Recently, neural architecture search (NAS) has been exploited to design feature pyramid networks (FPNs)
We propose a novel One-Shot Path Aggregation Network Architecture Search (OPANAS) algorithm, which significantly improves both searching efficiency and detection accuracy.
arXiv Detail & Related papers (2021-03-08T01:48:53Z) - Implicit Feature Pyramid Network for Object Detection [22.530998243247154]
We present an implicit feature pyramid network (i-FPN) for object detection.
We propose to use an implicit function, recently introduced in deep equilibrium model (DEQ) to model the transformation of FPN.
Experimental results on MS dataset show that i-FPN can significantly boost detection performance compared to baseline detectors.
arXiv Detail & Related papers (2020-12-25T11:30:27Z) - Dual Refinement Feature Pyramid Networks for Object Detection [2.88935873409577]
FPN is a common component used in object detectors, it supplements multi-scale information by adjacent level features and summation.
In this paper, we analyze the design defects from pixel level and feature map level.
We design a novel parameter-free feature pyramid networks named Dual Refinement Feature Pyramid Networks for the problems.
arXiv Detail & Related papers (2020-12-03T07:17:03Z) - NETNet: Neighbor Erasing and Transferring Network for Better Single Shot
Object Detection [170.30694322460045]
We propose a new Neighbor Erasing and Transferring (NET) mechanism to reconfigure the pyramid features and explore scale-aware features.
A single-shot network called NETNet is constructed for scale-aware object detection.
arXiv Detail & Related papers (2020-01-18T15:21:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.