Deep Co-supervision and Attention Fusion Strategy for Automatic COVID-19
Lung Infection Segmentation on CT Images
- URL: http://arxiv.org/abs/2112.10368v1
- Date: Mon, 20 Dec 2021 07:32:39 GMT
- Title: Deep Co-supervision and Attention Fusion Strategy for Automatic COVID-19
Lung Infection Segmentation on CT Images
- Authors: Haigen Hu, Leizhao Shen, Qiu Guan, Xiaoxin Li, Qianwei Zhou and Su
Ruan
- Abstract summary: In this paper, a novel segmentation scheme is proposed for the infections of COVID-19 on CT images.
A deep collaborative supervision scheme is proposed to guide the network learning the features of edges and semantics.
The effectiveness of the proposed scheme is demonstrated on four various COVID-19 CT datasets.
- Score: 1.898617934078969
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to the irregular shapes,various sizes and indistinguishable boundaries
between the normal and infected tissues, it is still a challenging task to
accurately segment the infected lesions of COVID-19 on CT images. In this
paper, a novel segmentation scheme is proposed for the infections of COVID-19
by enhancing supervised information and fusing multi-scale feature maps of
different levels based on the encoder-decoder architecture. To this end, a deep
collaborative supervision (Co-supervision) scheme is proposed to guide the
network learning the features of edges and semantics. More specifically, an
Edge Supervised Module (ESM) is firstly designed to highlight low-level
boundary features by incorporating the edge supervised information into the
initial stage of down-sampling. Meanwhile, an Auxiliary Semantic Supervised
Module (ASSM) is proposed to strengthen high-level semantic information by
integrating mask supervised information into the later stage. Then an Attention
Fusion Module (AFM) is developed to fuse multiple scale feature maps of
different levels by using an attention mechanism to reduce the semantic gaps
between high-level and low-level feature maps. Finally, the effectiveness of
the proposed scheme is demonstrated on four various COVID-19 CT datasets. The
results show that the proposed three modules are all promising. Based on the
baseline (ResUnet), using ESM, ASSM, or AFM alone can respectively increase
Dice metric by 1.12\%, 1.95\%,1.63\% in our dataset, while the integration by
incorporating three models together can rise 3.97\%. Compared with the existing
approaches in various datasets, the proposed method can obtain better
segmentation performance in some main metrics, and can achieve the best
generalization and comprehensive performance.
Related papers
- A Mutual Inclusion Mechanism for Precise Boundary Segmentation in Medical Images [2.9137615132901704]
We present a novel deep learning-based approach, MIPC-Net, for precise boundary segmentation in medical images.
We introduce the MIPC module, which enhances the focus on channel information when extracting position features.
We also propose the GL-MIPC-Residue, a global residual connection that enhances the integration of the encoder and decoder.
arXiv Detail & Related papers (2024-04-12T02:14:35Z) - SDR-Former: A Siamese Dual-Resolution Transformer for Liver Lesion
Classification Using 3D Multi-Phase Imaging [59.78761085714715]
This study proposes a novel Siamese Dual-Resolution Transformer (SDR-Former) framework for liver lesion classification.
The proposed framework has been validated through comprehensive experiments on two clinical datasets.
To support the scientific community, we are releasing our extensive multi-phase MR dataset for liver lesion analysis to the public.
arXiv Detail & Related papers (2024-02-27T06:32:56Z) - Generalizable Entity Grounding via Assistance of Large Language Model [77.07759442298666]
We propose a novel approach to densely ground visual entities from a long caption.
We leverage a large multimodal model to extract semantic nouns, a class-a segmentation model to generate entity-level segmentation, and a multi-modal feature fusion module to associate each semantic noun with its corresponding segmentation mask.
arXiv Detail & Related papers (2024-02-04T16:06:05Z) - Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - MLF-DET: Multi-Level Fusion for Cross-Modal 3D Object Detection [54.52102265418295]
We propose a novel and effective Multi-Level Fusion network, named as MLF-DET, for high-performance cross-modal 3D object DETection.
For the feature-level fusion, we present the Multi-scale Voxel Image fusion (MVI) module, which densely aligns multi-scale voxel features with image features.
For the decision-level fusion, we propose the lightweight Feature-cued Confidence Rectification (FCR) module, which exploits image semantics to rectify the confidence of detection candidates.
arXiv Detail & Related papers (2023-07-18T11:26:02Z) - Superresolution and Segmentation of OCT scans using Multi-Stage
adversarial Guided Attention Training [18.056525121226862]
We propose the multi-stage & multi-discriminatory generative adversarial network (MultiSDGAN) to translate OCT scans in high-resolution segmentation labels.
We evaluate and compare various combinations of channel and spatial attention to the MultiSDGAN architecture to extract more powerful feature maps.
Our results demonstrate relative improvements of 21.44% and 19.45% on the Dice coefficient and SSIM, respectively.
arXiv Detail & Related papers (2022-06-10T00:26:55Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - TC-Net: Triple Context Network for Automated Stroke Lesion Segmentation [0.5482532589225552]
We propose a new network, Triple Context Network (TC-Net), with the capture of spatial contextual information as the core.
Our network is evaluated on the open dataset ATLAS, achieving the highest score of 0.594, Hausdorff distance of 27.005 mm, and average symmetry surface distance of 7.137 mm.
arXiv Detail & Related papers (2022-02-28T11:12:16Z) - Deep Attention-guided Graph Clustering with Dual Self-supervision [49.040136530379094]
We propose a novel method, namely deep attention-guided graph clustering with dual self-supervision (DAGC)
We develop a dual self-supervision solution consisting of a soft self-supervision strategy with a triplet Kullback-Leibler divergence loss and a hard self-supervision strategy with a pseudo supervision loss.
Our method consistently outperforms state-of-the-art methods on six benchmark datasets.
arXiv Detail & Related papers (2021-11-10T06:53:03Z) - PAENet: A Progressive Attention-Enhanced Network for 3D to 2D Retinal
Vessel Segmentation [0.0]
3D to 2D retinal vessel segmentation is a challenging problem in Optical Coherence Tomography Angiography ( OCTA) images.
We propose a Progressive Attention-Enhanced Network (PAENet) based on attention mechanisms to extract rich feature representation.
Our proposed algorithm achieves state-of-the-art performance compared with previous methods.
arXiv Detail & Related papers (2021-08-26T10:27:25Z) - MPG-Net: Multi-Prediction Guided Network for Segmentation of Retinal
Layers in OCT Images [11.370735571629602]
We propose a novel multiprediction guided attention network (MPG-Net) for automated retinal layer segmentation in OCT images.
MPG-Net consists of two major steps to strengthen the discriminative power of a U-shape Fully convolutional network (FCN) for reliable automated segmentation.
arXiv Detail & Related papers (2020-09-28T21:22:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.