CAFCT: Contextual and Attentional Feature Fusions of Convolutional
Neural Networks and Transformer for Liver Tumor Segmentation
- URL: http://arxiv.org/abs/2401.16886v1
- Date: Tue, 30 Jan 2024 10:42:11 GMT
- Title: CAFCT: Contextual and Attentional Feature Fusions of Convolutional
Neural Networks and Transformer for Liver Tumor Segmentation
- Authors: Ming Kang, Chee-Ming Ting, Fung Fung Ting, Rapha\"el Phan
- Abstract summary: We propose a Contextual and Attentional feature Fusions enhanced Convolutional Network (CNN) and Transformer hybrid network (CAFCT) model for liver tumor segmentation.
Experimental results show that the proposed CAFCT achieves semantic Intersection of 90.38% and Dice score of 86.78%, respectively.
- Score: 4.255240258747643
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical image semantic segmentation techniques can help identify tumors
automatically from computed tomography (CT) scans. In this paper, we propose a
Contextual and Attentional feature Fusions enhanced Convolutional Neural
Network (CNN) and Transformer hybrid network (CAFCT) model for liver tumor
segmentation. In the proposed model, three other modules are introduced in the
network architecture: Attentional Feature Fusion (AFF), Atrous Spatial Pyramid
Pooling (ASPP) of DeepLabv3, and Attention Gates (AGs) to improve contextual
information related to tumor boundaries for accurate segmentation. Experimental
results show that the proposed CAFCT achieves a mean Intersection over Union
(IoU) of 90.38% and Dice score of 86.78%, respectively, on the Liver Tumor
Segmentation Benchmark (LiTS) dataset, outperforming pure CNN or Transformer
methods, e.g., Attention U-Net, and PVTFormer.
Related papers
- Attention-Enhanced Hybrid Feature Aggregation Network for 3D Brain Tumor Segmentation [0.9897828700959131]
Glioblastoma is a highly aggressive and malignant brain tumor type that requires early diagnosis and prompt intervention.
To address this challenge, Artificial Intelligence (AI)-driven approaches in healthcare have generated interest in efficiently diagnosing and evaluating brain tumors.
In our approach, we utilize a multi-scale, attention-guided and hybrid U-Net-shaped model -- GLIMS -- to perform 3D brain tumor segmentation.
arXiv Detail & Related papers (2024-03-15T00:52:17Z) - SDR-Former: A Siamese Dual-Resolution Transformer for Liver Lesion
Classification Using 3D Multi-Phase Imaging [59.78761085714715]
This study proposes a novel Siamese Dual-Resolution Transformer (SDR-Former) framework for liver lesion classification.
The proposed framework has been validated through comprehensive experiments on two clinical datasets.
To support the scientific community, we are releasing our extensive multi-phase MR dataset for liver lesion analysis to the public.
arXiv Detail & Related papers (2024-02-27T06:32:56Z) - Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network [63.845552349914186]
Capturing global contextual information plays a critical role in breast ultrasound (BUS) image classification.
Vision Transformers have an improved capability of capturing global contextual information but may distort the local image patterns due to the tokenization operations.
In this study, we proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN, designed to perform BUS tumor classification and segmentation.
arXiv Detail & Related papers (2023-08-04T01:19:32Z) - Reliable Joint Segmentation of Retinal Edema Lesions in OCT Images [55.83984261827332]
In this paper, we propose a novel reliable multi-scale wavelet-enhanced transformer network.
We develop a novel segmentation backbone that integrates a wavelet-enhanced feature extractor network and a multi-scale transformer module.
Our proposed method achieves better segmentation accuracy with a high degree of reliability as compared to other state-of-the-art segmentation approaches.
arXiv Detail & Related papers (2022-12-01T07:32:56Z) - Large-Kernel Attention for 3D Medical Image Segmentation [14.76728117630242]
In this paper, a novel large- kernel (LK) attention module is proposed to achieve accurate multi-organ segmentation and tumor segmentation.
The advantages of convolution and self-attention are combined in the proposed LK attention module, including local contextual information, long-range dependence, and channel adaptation.
The module also decomposes the LK convolution to optimize the computational cost and can be easily incorporated into FCNs such as U-Net.
arXiv Detail & Related papers (2022-07-19T16:32:55Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - A unified 3D framework for Organs at Risk Localization and Segmentation
for Radiation Therapy Planning [56.52933974838905]
Current medical workflow requires manual delineation of organs-at-risk (OAR)
In this work, we aim to introduce a unified 3D pipeline for OAR localization-segmentation.
Our proposed framework fully enables the exploitation of 3D context information inherent in medical imaging.
arXiv Detail & Related papers (2022-03-01T17:08:41Z) - Feature-enhanced Generation and Multi-modality Fusion based Deep Neural
Network for Brain Tumor Segmentation with Missing MR Modalities [2.867517731896504]
The main problem is that not all types of MRIs are always available in clinical exams.
We propose a novel brain tumor segmentation network in the case of missing one or more modalities.
The proposed network consists of three sub-networks: a feature-enhanced generator, a correlation constraint block and a segmentation network.
arXiv Detail & Related papers (2021-11-08T10:59:40Z) - BiTr-Unet: a CNN-Transformer Combined Network for MRI Brain Tumor
Segmentation [2.741266294612776]
We present a CNN-Transformer combined model called BiTr-Unet for brain tumor segmentation on multi-modal MRI scans.
The proposed BiTr-Unet achieves good performance on the BraTS 2021 validation dataset with mean Dice score 0.9076, 0.8392 and 0.8231, and mean Hausdorff distance 4.5322, 13.4592 and 14.9963 for the whole tumor, tumor core, and enhancing tumor, respectively.
arXiv Detail & Related papers (2021-09-25T04:18:34Z) - Automatic size and pose homogenization with spatial transformer network
to improve and accelerate pediatric segmentation [51.916106055115755]
We propose a new CNN architecture that is pose and scale invariant thanks to the use of Spatial Transformer Network (STN)
Our architecture is composed of three sequential modules that are estimated together during training.
We test the proposed method in kidney and renal tumor segmentation on abdominal pediatric CT scanners.
arXiv Detail & Related papers (2021-07-06T14:50:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.