Fully Attentional Network for Semantic Segmentation
- URL: http://arxiv.org/abs/2112.04108v1
- Date: Wed, 8 Dec 2021 04:34:55 GMT
- Title: Fully Attentional Network for Semantic Segmentation
- Authors: Qi Song, Jie Li, Chenghong Li, Hao Guo, Rui Huang
- Abstract summary: We propose Fully Attentional Network (FLANet) to encode both spatial and channel attentions in a single similarity map.
Our new method has achieved state-of-the-art performance on three challenging semantic segmentation datasets.
- Score: 17.24768249911501
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent non-local self-attention methods have proven to be effective in
capturing long-range dependencies for semantic segmentation. These methods
usually form a similarity map of RC*C (by compressing spatial dimensions) or
RHW*HW (by compressing channels) to describe the feature relations along either
channel or spatial dimensions, where C is the number of channels, H and W are
the spatial dimensions of the input feature map. However, such practices tend
to condense feature dependencies along the other dimensions,hence causing
attention missing, which might lead to inferior results for small/thin
categories or inconsistent segmentation inside large objects. To address this
problem, we propose anew approach, namely Fully Attentional Network (FLANet),to
encode both spatial and channel attentions in a single similarity map while
maintaining high computational efficiency. Specifically, for each channel map,
our FLANet can harvest feature responses from all other channel maps, and the
associated spatial positions as well, through a novel fully attentional module.
Our new method has achieved state-of-the-art performance on three challenging
semantic segmentation datasets,i.e., 83.6%, 46.99%, and 88.5% on the Cityscapes
test set,the ADE20K validation set, and the PASCAL VOC test set,respectively.
Related papers
- Efficient Multi-Scale Attention Module with Cross-Spatial Learning [4.046170185945849]
A novel efficient multi-scale attention (EMA) module is proposed.
We focus on retaining the information on per channel and decreasing the computational overhead.
We conduct extensive ablation studies and experiments on image classification and object detection tasks.
arXiv Detail & Related papers (2023-05-23T00:35:47Z) - UNETR++: Delving into Efficient and Accurate 3D Medical Image Segmentation [93.88170217725805]
We propose a 3D medical image segmentation approach, named UNETR++, that offers both high-quality segmentation masks as well as efficiency in terms of parameters, compute cost, and inference speed.
The core of our design is the introduction of a novel efficient paired attention (EPA) block that efficiently learns spatial and channel-wise discriminative features.
Our evaluations on five benchmarks, Synapse, BTCV, ACDC, BRaTs, and Decathlon-Lung, reveal the effectiveness of our contributions in terms of both efficiency and accuracy.
arXiv Detail & Related papers (2022-12-08T18:59:57Z) - Learning Implicit Feature Alignment Function for Semantic Segmentation [51.36809814890326]
Implicit Feature Alignment function (IFA) is inspired by the rapidly expanding topic of implicit neural representations.
We show that IFA implicitly aligns the feature maps at different levels and is capable of producing segmentation maps in arbitrary resolutions.
Our method can be combined with improvement on various architectures, and it achieves state-of-the-art accuracy trade-off on common benchmarks.
arXiv Detail & Related papers (2022-06-17T09:40:14Z) - TC-Net: Triple Context Network for Automated Stroke Lesion Segmentation [0.5482532589225552]
We propose a new network, Triple Context Network (TC-Net), with the capture of spatial contextual information as the core.
Our network is evaluated on the open dataset ATLAS, achieving the highest score of 0.594, Hausdorff distance of 27.005 mm, and average symmetry surface distance of 7.137 mm.
arXiv Detail & Related papers (2022-02-28T11:12:16Z) - Channelized Axial Attention for Semantic Segmentation [70.14921019774793]
We propose the Channelized Axial Attention (CAA) to seamlessly integratechannel attention and axial attention with reduced computationalcomplexity.
Our CAA not onlyrequires much less computation resources compared with otherdual attention models such as DANet, but also outperforms the state-of-the-art ResNet-101-based segmentation models on alltested datasets.
arXiv Detail & Related papers (2021-01-19T03:08:03Z) - Channel-wise Knowledge Distillation for Dense Prediction [73.99057249472735]
We propose to align features channel-wise between the student and teacher networks.
We consistently achieve superior performance on three benchmarks with various network structures.
arXiv Detail & Related papers (2020-11-26T12:00:38Z) - Dual Attention GANs for Semantic Image Synthesis [101.36015877815537]
We propose a novel Dual Attention GAN (DAGAN) to synthesize photo-realistic and semantically-consistent images.
We also propose two novel modules, i.e., position-wise Spatial Attention Module (SAM) and scale-wise Channel Attention Module (CAM)
DAGAN achieves remarkably better results than state-of-the-art methods, while using fewer model parameters.
arXiv Detail & Related papers (2020-08-29T17:49:01Z) - Adaptive feature recombination and recalibration for semantic
segmentation with Fully Convolutional Networks [57.64866581615309]
We propose recombination of features and a spatially adaptive recalibration block that is adapted for semantic segmentation with Fully Convolutional Networks.
Results indicate that Recombination and Recalibration improve the results of a competitive baseline, and generalize across three different problems.
arXiv Detail & Related papers (2020-06-19T15:45:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.