Deep Learning Framework with Multi-Head Dilated Encoders for Enhanced
Segmentation of Cervical Cancer on Multiparametric Magnetic Resonance Imaging
- URL: http://arxiv.org/abs/2306.11137v1
- Date: Mon, 19 Jun 2023 19:41:21 GMT
- Title: Deep Learning Framework with Multi-Head Dilated Encoders for Enhanced
Segmentation of Cervical Cancer on Multiparametric Magnetic Resonance Imaging
- Authors: Reza Kalantar, Sebastian Curcean, Jessica M Winfield, Gigin Lin,
Christina Messiou, Matthew D Blackledge and Dow-Mu Koh
- Abstract summary: T2-weighted magnetic resonance imaging (MRI) and diffusion-weighted imaging (DWI) are essential components for cervical cancer diagnosis.
We propose a novel multi-head framework that uses dilated convolutions and shared residual connections for separate encoding of multiparametric MRI images.
- Score: 0.6597195879147557
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: T2-weighted magnetic resonance imaging (MRI) and diffusion-weighted imaging
(DWI) are essential components for cervical cancer diagnosis. However,
combining these channels for training deep learning models are challenging due
to misalignment of images. Here, we propose a novel multi-head framework that
uses dilated convolutions and shared residual connections for separate encoding
of multiparametric MRI images. We employ a residual U-Net model as a baseline,
and perform a series of architectural experiments to evaluate the tumor
segmentation performance based on multiparametric input channels and feature
encoding configurations. All experiments were performed using a cohort
including 207 patients with locally advanced cervical cancer. Our proposed
multi-head model using separate dilated encoding for T2W MRI, and combined
b1000 DWI and apparent diffusion coefficient (ADC) images achieved the best
median Dice coefficient similarity (DSC) score, 0.823 (95% confidence interval
(CI), 0.595-0.797), outperforming the conventional multi-channel model, DSC
0.788 (95% CI, 0.568-0.776), although the difference was not statistically
significant (p>0.05). We investigated channel sensitivity using 3D GRAD-CAM and
channel dropout, and highlighted the critical importance of T2W and ADC
channels for accurate tumor segmentations. However, our results showed that
b1000 DWI had a minor impact on overall segmentation performance. We
demonstrated that the use of separate dilated feature extractors and
independent contextual learning improved the model's ability to reduce the
boundary effects and distortion of DWI, leading to improved segmentation
performance. Our findings can have significant implications for the development
of robust and generalizable models that can extend to other multi-modal
segmentation applications.
Related papers
- Multi-Layer Feature Fusion with Cross-Channel Attention-Based U-Net for Kidney Tumor Segmentation [0.0]
U-Net based deep learning techniques are emerging as a promising approach for automated medical image segmentation.
We present an improved U-Net based model for end-to-end automated semantic segmentation of CT scan images to identify renal tumors.
arXiv Detail & Related papers (2024-10-20T19:02:41Z) - Unifying Subsampling Pattern Variations for Compressed Sensing MRI with Neural Operators [72.79532467687427]
Compressed Sensing MRI reconstructs images of the body's internal anatomy from undersampled and compressed measurements.
Deep neural networks have shown great potential for reconstructing high-quality images from highly undersampled measurements.
We propose a unified model that is robust to different subsampling patterns and image resolutions in CS-MRI.
arXiv Detail & Related papers (2024-10-05T20:03:57Z) - TotalSegmentator MRI: Sequence-Independent Segmentation of 59 Anatomical Structures in MR images [62.53931644063323]
In this study we extended the capabilities of TotalSegmentator to MR images.
We trained an nnU-Net segmentation algorithm on this dataset and calculated similarity coefficients (Dice) to evaluate the model's performance.
The model significantly outperformed two other publicly available segmentation models (Dice score 0.824 versus 0.762; p0.001 and 0.762 versus 0.542; p)
arXiv Detail & Related papers (2024-05-29T20:15:54Z) - MRSegmentator: Robust Multi-Modality Segmentation of 40 Classes in MRI and CT Sequences [4.000329151950926]
The model was trained on 1,200 manually annotated MRI scans from the UK Biobank, 221 in-house MRI scans and 1228 CT scans.
It showcased high accuracy in segmenting well-defined organs, achieving Dice Similarity Coefficient (DSC) scores of 0.97 for the right and left lungs, and 0.95 for the heart.
It also demonstrated robustness in organs like the liver (DSC: 0.96) and kidneys (DSC: 0.95 left, 0.95 right), which present more variability.
arXiv Detail & Related papers (2024-05-10T13:15:42Z) - 3D Lymphoma Segmentation on PET/CT Images via Multi-Scale Information Fusion with Cross-Attention [6.499725732124126]
This study aims to develop a precise segmentation method for diffuse large B-cell lymphoma (DLBCL) lesions.
We propose a 3D dual-branch encoder segmentation method using shifted window transformers and a Multi-Scale Information Fusion (MSIF) module.
The model was trained and validated on a dataset of 165 DLBCL patients using 5-fold cross-validation.
arXiv Detail & Related papers (2024-02-04T05:25:12Z) - Cross-modality Guidance-aided Multi-modal Learning with Dual Attention
for MRI Brain Tumor Grading [47.50733518140625]
Brain tumor represents one of the most fatal cancers around the world, and is very common in children and the elderly.
We propose a novel cross-modality guidance-aided multi-modal learning with dual attention for addressing the task of MRI brain tumor grading.
arXiv Detail & Related papers (2024-01-17T07:54:49Z) - A Two-Stage Generative Model with CycleGAN and Joint Diffusion for
MRI-based Brain Tumor Detection [41.454028276986946]
We propose a novel framework Two-Stage Generative Model (TSGM) to improve brain tumor detection and segmentation.
CycleGAN is trained on unpaired data to generate abnormal images from healthy images as data prior.
VE-JP is implemented to reconstruct healthy images using synthetic paired abnormal images as a guide.
arXiv Detail & Related papers (2023-11-06T12:58:26Z) - On Sensitivity and Robustness of Normalization Schemes to Input
Distribution Shifts in Automatic MR Image Diagnosis [58.634791552376235]
Deep Learning (DL) models have achieved state-of-the-art performance in diagnosing multiple diseases using reconstructed images as input.
DL models are sensitive to varying artifacts as it leads to changes in the input data distribution between the training and testing phases.
We propose to use other normalization techniques, such as Group Normalization and Layer Normalization, to inject robustness into model performance against varying image artifacts.
arXiv Detail & Related papers (2023-06-23T03:09:03Z) - Reliable Joint Segmentation of Retinal Edema Lesions in OCT Images [55.83984261827332]
In this paper, we propose a novel reliable multi-scale wavelet-enhanced transformer network.
We develop a novel segmentation backbone that integrates a wavelet-enhanced feature extractor network and a multi-scale transformer module.
Our proposed method achieves better segmentation accuracy with a high degree of reliability as compared to other state-of-the-art segmentation approaches.
arXiv Detail & Related papers (2022-12-01T07:32:56Z) - Affinity Feature Strengthening for Accurate, Complete and Robust Vessel
Segmentation [48.638327652506284]
Vessel segmentation is crucial in many medical image applications, such as detecting coronary stenoses, retinal vessel diseases and brain aneurysms.
We present a novel approach, the affinity feature strengthening network (AFN), which jointly models geometry and refines pixel-wise segmentation features using a contrast-insensitive, multiscale affinity approach.
arXiv Detail & Related papers (2022-11-12T05:39:17Z) - PA-ResSeg: A Phase Attention Residual Network for Liver Tumor
Segmentation from Multi-phase CT Images [19.725599681891925]
We propose a phase attention residual network (PA-ResSeg) to model multi-phase features for accurate liver tumor segmentation.
The proposed method shows its robustness and generalization capability in different datasets and different backbones.
arXiv Detail & Related papers (2021-02-27T17:30:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.