Hyper-Connected Transformer Network for Multi-Modality PET-CT
Segmentation
- URL: http://arxiv.org/abs/2210.15808v2
- Date: Mon, 7 Aug 2023 10:33:34 GMT
- Title: Hyper-Connected Transformer Network for Multi-Modality PET-CT
Segmentation
- Authors: Lei Bi, Michael Fulham, Shaoli Song, David Dagan Feng, Jinman Kim
- Abstract summary: Co-learning complementary PET-CT imaging features is a fundamental requirement for automatic tumor segmentation.
We propose a hyper-connected transformer network that integrates a transformer network (TN) with a hyper connected fusion for multi-modality PET-CT images.
Our results with two clinical datasets show that HCT achieved better performance in segmentation accuracy when compared to the existing methods.
- Score: 16.999643199612244
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: [18F]-Fluorodeoxyglucose (FDG) positron emission tomography - computed
tomography (PET-CT) has become the imaging modality of choice for diagnosing
many cancers. Co-learning complementary PET-CT imaging features is a
fundamental requirement for automatic tumor segmentation and for developing
computer aided cancer diagnosis systems. In this study, we propose a
hyper-connected transformer (HCT) network that integrates a transformer network
(TN) with a hyper connected fusion for multi-modality PET-CT images. The TN was
leveraged for its ability to provide global dependencies in image feature
learning, which was achieved by using image patch embeddings with a
self-attention mechanism to capture image-wide contextual information. We
extended the single-modality definition of TN with multiple TN based branches
to separately extract image features. We also introduced a hyper connected
fusion to fuse the contextual and complementary image features across multiple
transformers in an iterative manner. Our results with two clinical datasets
show that HCT achieved better performance in segmentation accuracy when
compared to the existing methods.
Related papers
- DCT-HistoTransformer: Efficient Lightweight Vision Transformer with DCT Integration for histopathological image analysis [0.0]
We introduce a novel lightweight breast cancer classification approach using Vision Transformers (ViTs)
By incorporating parallel processing pathways for Discrete Cosine Transform (DCT) Attention and MobileConv, we convert image data from the spatial domain to the frequency domain to utilize the benefits such as filtering out high frequencies in the image.
Our proposed model achieves an accuracy of 96.00% $pm$ 0.48% for binary classification and 87.85% $pm$ 0.93% for multiclass classification, which is comparable to state-of-the-art models.
arXiv Detail & Related papers (2024-10-24T21:16:56Z) - Prototype Learning Guided Hybrid Network for Breast Tumor Segmentation in DCE-MRI [58.809276442508256]
We propose a hybrid network via the combination of convolution neural network (CNN) and transformer layers.
The experimental results on private and public DCE-MRI datasets demonstrate that the proposed hybrid network superior performance than the state-of-the-art methods.
arXiv Detail & Related papers (2024-08-11T15:46:00Z) - Multi-modal Evidential Fusion Network for Trusted PET/CT Tumor Segmentation [5.839660501978193]
The quality of PET and CT images varies widely in clinical settings, which leads to uncertainty in the modality information extracted by networks.
This paper proposes a novel Multi-modal Evidential Fusion Network (MEFN) comprising a Cross-Modal Feature Learning (CFL) module and a Multi-modal Trusted Fusion (MTF) module.
Our model can provide radiologists with credible uncertainty of the segmentation results for their decision in accepting or rejecting the automatic segmentation results.
arXiv Detail & Related papers (2024-06-26T13:14:24Z) - SDR-Former: A Siamese Dual-Resolution Transformer for Liver Lesion
Classification Using 3D Multi-Phase Imaging [59.78761085714715]
This study proposes a novel Siamese Dual-Resolution Transformer (SDR-Former) framework for liver lesion classification.
The proposed framework has been validated through comprehensive experiments on two clinical datasets.
To support the scientific community, we are releasing our extensive multi-phase MR dataset for liver lesion analysis to the public.
arXiv Detail & Related papers (2024-02-27T06:32:56Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network [63.845552349914186]
Capturing global contextual information plays a critical role in breast ultrasound (BUS) image classification.
Vision Transformers have an improved capability of capturing global contextual information but may distort the local image patterns due to the tokenization operations.
In this study, we proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN, designed to perform BUS tumor classification and segmentation.
arXiv Detail & Related papers (2023-08-04T01:19:32Z) - SwinCross: Cross-modal Swin Transformer for Head-and-Neck Tumor
Segmentation in PET/CT Images [6.936329289469511]
Cross-Modal Swin Transformer (SwinCross) with cross-modal attention (CMA) module incorporated cross-modal feature extraction at multiple resolutions.
The proposed method is experimentally shown to outperform state-of-the-art transformer-based methods.
arXiv Detail & Related papers (2023-02-08T03:36:57Z) - CyTran: A Cycle-Consistent Transformer with Multi-Level Consistency for
Non-Contrast to Contrast CT Translation [56.622832383316215]
We propose a novel approach to translate unpaired contrast computed tomography (CT) scans to non-contrast CT scans.
Our approach is based on cycle-consistent generative adversarial convolutional transformers, for short, CyTran.
Our empirical results show that CyTran outperforms all competing methods.
arXiv Detail & Related papers (2021-10-12T23:25:03Z) - ResViT: Residual vision transformers for multi-modal medical image
synthesis [0.0]
We propose a novel generative adversarial approach for medical image synthesis, ResViT, to combine local precision of convolution operators with contextual sensitivity of vision transformers.
Our results indicate the superiority of ResViT against competing methods in terms of qualitative observations and quantitative metrics.
arXiv Detail & Related papers (2021-06-30T12:57:37Z) - TransMed: Transformers Advance Multi-modal Medical Image Classification [4.500880052705654]
convolutional neural networks (CNN) have shown very competitive performance in medical image analysis tasks.
Transformers have been applied to computer vision and achieved remarkable success in large-scale datasets.
TransMed combines the advantages of CNN and transformer to efficiently extract low-level features of images.
arXiv Detail & Related papers (2021-03-10T08:57:53Z) - A Multi-Stage Attentive Transfer Learning Framework for Improving
COVID-19 Diagnosis [49.3704402041314]
We propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis.
Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains.
Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images.
arXiv Detail & Related papers (2021-01-14T01:39:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.