SDR-Former: A Siamese Dual-Resolution Transformer for Liver Lesion
Classification Using 3D Multi-Phase Imaging
- URL: http://arxiv.org/abs/2402.17246v1
- Date: Tue, 27 Feb 2024 06:32:56 GMT
- Title: SDR-Former: A Siamese Dual-Resolution Transformer for Liver Lesion
Classification Using 3D Multi-Phase Imaging
- Authors: Meng Lou, Hanning Ying, Xiaoqing Liu, Hong-Yu Zhou, Yuqing Zhang,
Yizhou Yu
- Abstract summary: This study proposes a novel Siamese Dual-Resolution Transformer (SDR-Former) framework for liver lesion classification.
The proposed framework has been validated through comprehensive experiments on two clinical datasets.
To support the scientific community, we are releasing our extensive multi-phase MR dataset for liver lesion analysis to the public.
- Score: 59.78761085714715
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automated classification of liver lesions in multi-phase CT and MR scans is
of clinical significance but challenging. This study proposes a novel Siamese
Dual-Resolution Transformer (SDR-Former) framework, specifically designed for
liver lesion classification in 3D multi-phase CT and MR imaging with varying
phase counts. The proposed SDR-Former utilizes a streamlined Siamese Neural
Network (SNN) to process multi-phase imaging inputs, possessing robust feature
representations while maintaining computational efficiency. The weight-sharing
feature of the SNN is further enriched by a hybrid Dual-Resolution Transformer
(DR-Former), comprising a 3D Convolutional Neural Network (CNN) and a tailored
3D Transformer for processing high- and low-resolution images, respectively.
This hybrid sub-architecture excels in capturing detailed local features and
understanding global contextual information, thereby, boosting the SNN's
feature extraction capabilities. Additionally, a novel Adaptive Phase Selection
Module (APSM) is introduced, promoting phase-specific intercommunication and
dynamically adjusting each phase's influence on the diagnostic outcome. The
proposed SDR-Former framework has been validated through comprehensive
experiments on two clinical datasets: a three-phase CT dataset and an
eight-phase MR dataset. The experimental results affirm the efficacy of the
proposed framework. To support the scientific community, we are releasing our
extensive multi-phase MR dataset for liver lesion analysis to the public. This
pioneering dataset, being the first publicly available multi-phase MR dataset
in this field, also underpins the MICCAI LLD-MMRI Challenge. The dataset is
accessible at:https://bit.ly/3IyYlgN.
Related papers
- Prototype Learning Guided Hybrid Network for Breast Tumor Segmentation in DCE-MRI [58.809276442508256]
We propose a hybrid network via the combination of convolution neural network (CNN) and transformer layers.
The experimental results on private and public DCE-MRI datasets demonstrate that the proposed hybrid network superior performance than the state-of-the-art methods.
arXiv Detail & Related papers (2024-08-11T15:46:00Z) - BTDNet: a Multi-Modal Approach for Brain Tumor Radiogenomic
Classification [14.547418131610188]
This paper proposes a novel multi-modal approach, BTDNet, to predict MGMT promoter methylation status.
The proposed method outperforms by large margins the state-of-the-art methods in the RSNA-ASNR-MICCAI BraTS 2021 Challenge.
arXiv Detail & Related papers (2023-10-05T11:56:06Z) - 3DSAM-adapter: Holistic adaptation of SAM from 2D to 3D for promptable tumor segmentation [52.699139151447945]
We propose a novel adaptation method for transferring the segment anything model (SAM) from 2D to 3D for promptable medical image segmentation.
Our model can outperform domain state-of-the-art medical image segmentation models on 3 out of 4 tasks, specifically by 8.25%, 29.87%, and 10.11% for kidney tumor, pancreas tumor, colon cancer segmentation, and achieve similar performance for liver tumor segmentation.
arXiv Detail & Related papers (2023-06-23T12:09:52Z) - Multiscale Metamorphic VAE for 3D Brain MRI Synthesis [5.060516201839319]
Generative modeling of 3D brain MRIs presents difficulties in achieving high visual fidelity while ensuring sufficient coverage of the data distribution.
In this work, we propose to address this challenge with composable, multiscale morphological transformations in a variational autoencoder framework.
We show substantial performance improvements in FID while retaining comparable, or superior, reconstruction quality compared to prior work based on VAEs and generative adversarial networks (GANs)
arXiv Detail & Related papers (2023-01-09T09:15:30Z) - View-Disentangled Transformer for Brain Lesion Detection [50.4918615815066]
We propose a novel view-disentangled transformer to enhance the extraction of MRI features for more accurate tumour detection.
First, the proposed transformer harvests long-range correlation among different positions in a 3D brain scan.
Second, the transformer models a stack of slice features as multiple 2D views and enhance these features view-by-view.
Third, we deploy the proposed transformer module in a transformer backbone, which can effectively detect the 2D regions surrounding brain lesions.
arXiv Detail & Related papers (2022-09-20T11:58:23Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Semi-Supervised Hybrid Spine Network for Segmentation of Spine MR Images [14.190504802866288]
We propose a two-stage algorithm, named semi-supervised hybrid spine network (SSHSNet) to achieve simultaneous vertebral bodies (VBs) and intervertebral discs (IVDs) segmentation.
In the first stage, we constructed a 2D semi-supervised DeepLabv3+ by using cross pseudo supervision to obtain intra-slice features and coarse segmentation.
In the second stage, a 3D full-resolution patch-based DeepLabv3+ was built to extract inter-slice information.
Results show that the proposed method has great potential in dealing with the data imbalance problem
arXiv Detail & Related papers (2022-03-23T02:57:14Z) - Mutual Attention-based Hybrid Dimensional Network for Multimodal Imaging
Computer-aided Diagnosis [4.657804635843888]
We propose a novel mutual attention-based hybrid dimensional network for MultiModal 3D medical image classification (MMNet)
The hybrid dimensional network integrates 2D CNN with 3D convolution modules to generate deeper and more informative feature maps.
We further design a mutual attention framework in the network to build the region-wise consistency in similar stereoscopic regions of different image modalities.
arXiv Detail & Related papers (2022-01-24T02:31:25Z) - Swin UNETR: Swin Transformers for Semantic Segmentation of Brain Tumors
in MRI Images [7.334185314342017]
We propose a novel segmentation model termed Swin UNEt TRansformers (Swin UNETR)
The model extracts features at five different resolutions by utilizing shifted windows for computing self-attention.
We have participated in BraTS 2021 segmentation challenge, and our proposed model ranks among the top-performing approaches in the validation phase.
arXiv Detail & Related papers (2022-01-04T18:01:34Z) - MS-Net: Multi-Site Network for Improving Prostate Segmentation with
Heterogeneous MRI Data [75.73881040581767]
We propose a novel multi-site network (MS-Net) for improving prostate segmentation by learning robust representations.
Our MS-Net improves the performance across all datasets consistently, and outperforms state-of-the-art methods for multi-site learning.
arXiv Detail & Related papers (2020-02-09T14:11:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.