DPNET: Dual-Path Network for Efficient Object Detectioj with Lightweight
Self-Attention
- URL: http://arxiv.org/abs/2111.00500v1
- Date: Sun, 31 Oct 2021 13:38:16 GMT
- Title: DPNET: Dual-Path Network for Efficient Object Detectioj with Lightweight
Self-Attention
- Authors: Huimin Shi, Quan Zhou, Yinghao Ni, Xiaofu Wu and Longin Jan Latecki
- Abstract summary: DPNet is a dual path network for efficient object detection with lightweight self-attention.
It achieves 29.0% AP on COCO dataset, with only 1.14 GFLOPs and 2.27M model size for a 320x320 image.
- Score: 16.13989397708127
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object detection often costs a considerable amount of computation to get
satisfied performance, which is unfriendly to be deployed in edge devices. To
address the trade-off between computational cost and detection accuracy, this
paper presents a dual path network, named DPNet, for efficient object detection
with lightweight self-attention. In backbone, a single input/output lightweight
self-attention module (LSAM) is designed to encode global interactions between
different positions. LSAM is also extended into a multiple-inputs version in
feature pyramid network (FPN), which is employed to capture cross-resolution
dependencies in two paths. Extensive experiments on the COCO dataset
demonstrate that our method achieves state-of-the-art detection results. More
specifically, DPNet obtains 29.0% AP on COCO test-dev, with only 1.14 GFLOPs
and 2.27M model size for a 320x320 image.
Related papers
- BAFNet: Bilateral Attention Fusion Network for Lightweight Semantic Segmentation of Urban Remote Sensing Images [6.153725909241752]
We propose a lightweight semantic segmentation network called bilateral attention fusion network (BAFNet) to efficiently segment high-resolution urban remote sensing images.
BAFNet outperforms advanced lightweight models in accuracy but also demonstrates comparable performance to non-lightweight state-of-the-art methods on two datasets.
arXiv Detail & Related papers (2024-09-16T13:25:42Z) - Lightweight Salient Object Detection in Optical Remote-Sensing Images
via Semantic Matching and Edge Alignment [61.45639694373033]
We propose a novel lightweight network for optical remote sensing images (ORSI-SOD) based on semantic matching and edge alignment, termed SeaNet.
Specifically, SeaNet includes a lightweight MobileNet-V2 for feature extraction, a dynamic semantic matching module (DSMM) for high-level features, and a portable decoder for inference.
arXiv Detail & Related papers (2023-01-07T04:33:51Z) - DPNet: Dual-Path Network for Real-time Object Detection with Lightweight
Attention [15.360769793764526]
This paper presents a dual-path network, named DPNet, with a lightweight attention scheme for real-time object detection.
DPNet achieves state-of-the-art trade-off between detection accuracy and implementation efficiency.
arXiv Detail & Related papers (2022-09-28T09:11:01Z) - SALISA: Saliency-based Input Sampling for Efficient Video Object
Detection [58.22508131162269]
We propose SALISA, a novel non-uniform SALiency-based Input SAmpling technique for video object detection.
We show that SALISA significantly improves the detection of small objects.
arXiv Detail & Related papers (2022-04-05T17:59:51Z) - Lightweight Salient Object Detection in Optical Remote Sensing Images
via Feature Correlation [93.80710126516405]
We propose a novel lightweight ORSI-SOD solution, named CorrNet, to address these issues.
By reducing the parameters and computations of each component, CorrNet ends up having only 4.09M parameters and running with 21.09G FLOPs.
Experimental results on two public datasets demonstrate that our lightweight CorrNet achieves competitive or even better performance compared with 26 state-of-the-art methods.
arXiv Detail & Related papers (2022-01-20T08:28:01Z) - Anchor Retouching via Model Interaction for Robust Object Detection in
Aerial Images [15.404024559652534]
We present an effective Dynamic Enhancement Anchor (DEA) network to construct a novel training sample generator.
Our method achieves state-of-the-art performance in accuracy with moderate inference speed and computational overhead for training.
arXiv Detail & Related papers (2021-12-13T14:37:20Z) - EPMF: Efficient Perception-aware Multi-sensor Fusion for 3D Semantic Segmentation [62.210091681352914]
We study multi-sensor fusion for 3D semantic segmentation for many applications, such as autonomous driving and robotics.
In this work, we investigate a collaborative fusion scheme called perception-aware multi-sensor fusion (PMF)
We propose a two-stream network to extract features from the two modalities separately. The extracted features are fused by effective residual-based fusion modules.
arXiv Detail & Related papers (2021-06-21T10:47:26Z) - Dense Dual-Path Network for Real-time Semantic Segmentation [7.8381744043673045]
We introduce a novel Dual-Path Network (DDPNet) for real-time semantic segmentation under resource constraints.
DDPNet achieves 75.3% mIoU with 52.6 FPS for an input of 1024 X 2048 resolution on a single GTX 1080Ti card.
arXiv Detail & Related papers (2020-10-21T06:11:41Z) - A Single Stream Network for Robust and Real-time RGB-D Salient Object
Detection [89.88222217065858]
We design a single stream network to use the depth map to guide early fusion and middle fusion between RGB and depth.
This model is 55.5% lighter than the current lightest model and runs at a real-time speed of 32 FPS when processing a $384 times 384$ image.
arXiv Detail & Related papers (2020-07-14T04:40:14Z) - Binary DAD-Net: Binarized Driveable Area Detection Network for
Autonomous Driving [94.40107679615618]
This paper proposes a novel binarized driveable area detection network (binary DAD-Net)
It uses only binary weights and activations in the encoder, the bottleneck, and the decoder part.
It outperforms state-of-the-art semantic segmentation networks on public datasets.
arXiv Detail & Related papers (2020-06-15T07:09:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.