MA-Unet: An improved version of Unet based on multi-scale and attention
mechanism for medical image segmentation
- URL: http://arxiv.org/abs/2012.10952v1
- Date: Sun, 20 Dec 2020 15:29:18 GMT
- Title: MA-Unet: An improved version of Unet based on multi-scale and attention
mechanism for medical image segmentation
- Authors: Yutong Cai, Yong Wang
- Abstract summary: convolutional neural networks (CNNs) are promoting the development of medical image semantic segmentation.
In this paper, we try to eliminate semantic ambiguity in skip connection operations by adding attention gates (AGs)
Our model obtains better segmentation performance while introducing fewer parameters.
- Score: 4.082245106486775
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although convolutional neural networks (CNNs) are promoting the development
of medical image semantic segmentation, the standard model still has some
shortcomings. First, the feature mapping from the encoder and decoder
sub-networks in the skip connection operation has a large semantic difference.
Second, the remote feature dependence is not effectively modeled. Third, the
global context information of different scales is ignored. In this paper, we
try to eliminate semantic ambiguity in skip connection operations by adding
attention gates (AGs), and use attention mechanisms to combine local features
with their corresponding global dependencies, explicitly model the dependencies
between channels and use multi-scale predictive fusion to utilize global
information at different scales. Compared with other state-of-the-art
segmentation networks, our model obtains better segmentation performance while
introducing fewer parameters.
Related papers
- Skin Lesion Segmentation Improved by Transformer-based Networks with
Inter-scale Dependency Modeling [0.0]
Melanoma is a dangerous type of skin cancer resulting from abnormal skin cell growth.
The symmetrical U-Net model's reliance on convolutional operations hinders its ability to capture long-range dependencies.
Several Transformer-based U-Net topologies have recently been created to overcome this limitation.
arXiv Detail & Related papers (2023-10-20T15:53:51Z) - A Generic Shared Attention Mechanism for Various Backbone Neural Networks [53.36677373145012]
Self-attention modules (SAMs) produce strongly correlated attention maps across different layers.
Dense-and-Implicit Attention (DIA) shares SAMs across layers and employs a long short-term memory module.
Our simple yet effective DIA can consistently enhance various network backbones.
arXiv Detail & Related papers (2022-10-27T13:24:08Z) - Multi-scale and Cross-scale Contrastive Learning for Semantic
Segmentation [5.281694565226513]
We apply contrastive learning to enhance the discriminative power of the multi-scale features extracted by semantic segmentation networks.
By first mapping the encoder's multi-scale representations to a common feature space, we instantiate a novel form of supervised local-global constraint.
arXiv Detail & Related papers (2022-03-25T01:24:24Z) - Learning Target-aware Representation for Visual Tracking via Informative
Interactions [49.552877881662475]
We introduce a novel backbone architecture to improve target-perception ability of feature representation for tracking.
The proposed GIM module and InBN mechanism are general and applicable to different backbone types including CNN and Transformer.
arXiv Detail & Related papers (2022-01-07T16:22:27Z) - A Discriminative Channel Diversification Network for Image
Classification [21.049734250642974]
We propose a light-weight and effective attention module, called channel diversification block, to enhance the global context.
Unlike other channel attention mechanisms, the proposed module focuses on the most discriminative features.
Experiments on CIFAR-10, SVHN, and Tiny-ImageNet datasets demonstrate that the proposed module improves the performance of the baseline networks by a margin of 3% on average.
arXiv Detail & Related papers (2021-12-10T23:00:53Z) - Domain Attention Consistency for Multi-Source Domain Adaptation [100.25573559447551]
Key design is a feature channel attention module, which aims to identify transferable features (attributes)
Experiments on three MSDA benchmarks show that our DAC-Net achieves new state of the art performance on all of them.
arXiv Detail & Related papers (2021-11-06T15:56:53Z) - Encoder Fusion Network with Co-Attention Embedding for Referring Image
Segmentation [87.01669173673288]
We propose an encoder fusion network (EFN), which transforms the visual encoder into a multi-modal feature learning network.
A co-attention mechanism is embedded in the EFN to realize the parallel update of multi-modal features.
The experiment results on four benchmark datasets demonstrate that the proposed approach achieves the state-of-the-art performance without any post-processing.
arXiv Detail & Related papers (2021-05-05T02:27:25Z) - Channelized Axial Attention for Semantic Segmentation [70.14921019774793]
We propose the Channelized Axial Attention (CAA) to seamlessly integratechannel attention and axial attention with reduced computationalcomplexity.
Our CAA not onlyrequires much less computation resources compared with otherdual attention models such as DANet, but also outperforms the state-of-the-art ResNet-101-based segmentation models on alltested datasets.
arXiv Detail & Related papers (2021-01-19T03:08:03Z) - Multi-Attention-Network for Semantic Segmentation of Fine Resolution
Remote Sensing Images [10.835342317692884]
The accuracy of semantic segmentation in remote sensing images has been increased significantly by deep convolutional neural networks.
This paper proposes a Multi-Attention-Network (MANet) to address these issues.
A novel attention mechanism of kernel attention with linear complexity is proposed to alleviate the large computational demand in attention.
arXiv Detail & Related papers (2020-09-03T09:08:02Z) - Universal-RCNN: Universal Object Detector via Transferable Graph R-CNN [117.80737222754306]
We present a novel universal object detector called Universal-RCNN.
We first generate a global semantic pool by integrating all high-level semantic representation of all the categories.
An Intra-Domain Reasoning Module learns and propagates the sparse graph representation within one dataset guided by a spatial-aware GCN.
arXiv Detail & Related papers (2020-02-18T07:57:45Z) - Hybrid Multiple Attention Network for Semantic Segmentation in Aerial
Images [24.35779077001839]
We propose a novel attention-based framework named Hybrid Multiple Attention Network (HMANet) to adaptively capture global correlations.
We introduce a simple yet effective region shuffle attention (RSA) module to reduce feature redundant and improve the efficiency of self-attention mechanism.
arXiv Detail & Related papers (2020-01-09T07:47:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.