Generalized Dice Focal Loss trained 3D Residual UNet for Automated
Lesion Segmentation in Whole-Body FDG PET/CT Images
- URL: http://arxiv.org/abs/2309.13553v1
- Date: Sun, 24 Sep 2023 05:29:45 GMT
- Title: Generalized Dice Focal Loss trained 3D Residual UNet for Automated
Lesion Segmentation in Whole-Body FDG PET/CT Images
- Authors: Shadab Ahamed, Arman Rahmim
- Abstract summary: We train a 3D Residual UNet using Generalized Dice Focal Loss function on the AutoPET challenge 2023 training dataset.
On the preliminary test phase, the average ensemble achieved a Dice similarity coefficient (DSC), false-positive volume (FPV) and false negative volume (FNV) of 0.5417, 0.8261 ml, and 0.2538 ml, respectively.
- Score: 0.4630436098920747
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automated segmentation of cancerous lesions in PET/CT images is a vital
initial task for quantitative analysis. However, it is often challenging to
train deep learning-based segmentation methods to high degree of accuracy due
to the diversity of lesions in terms of their shapes, sizes, and radiotracer
uptake levels. These lesions can be found in various parts of the body, often
close to healthy organs that also show significant uptake. Consequently,
developing a comprehensive PET/CT lesion segmentation model is a demanding
endeavor for routine quantitative image analysis. In this work, we train a 3D
Residual UNet using Generalized Dice Focal Loss function on the AutoPET
challenge 2023 training dataset. We develop our models in a 5-fold
cross-validation setting and ensemble the five models via average and
weighted-average ensembling. On the preliminary test phase, the average
ensemble achieved a Dice similarity coefficient (DSC), false-positive volume
(FPV) and false negative volume (FNV) of 0.5417, 0.8261 ml, and 0.2538 ml,
respectively, while the weighted-average ensemble achieved 0.5417, 0.8186 ml,
and 0.2538 ml, respectively. Our algorithm can be accessed via this link:
https://github.com/ahxmeds/autosegnet.
Related papers
- AutoPET Challenge III: Testing the Robustness of Generalized Dice Focal Loss trained 3D Residual UNet for FDG and PSMA Lesion Segmentation from Whole-Body PET/CT Images [0.0]
In this study, we utilize a 3D Residual UNet model and employ the Generalized Dice Loss function to train the model on the AutoPET Challenge 2024 dataset.
In the preliminary test phase for Task-1, the average ensemble achieved a mean Dice Similarity Coefficient (DSC) of 0.6687, mean false negative volume (FNV) of 10.9522 ml and mean false positive volume (FPV) 2.9684 ml.
arXiv Detail & Related papers (2024-09-16T10:27:30Z) - From FDG to PSMA: A Hitchhiker's Guide to Multitracer, Multicenter Lesion Segmentation in PET/CT Imaging [0.9384264274298444]
We present our solution for the autoPET III challenge, targeting multitracer, multicenter generalization using the nnU-Net framework with the ResEncL architecture.
Key techniques include misalignment data augmentation and multi-modal pretraining across CT, MR, and PET datasets.
Compared to the default nnU-Net, which achieved a Dice score of 57.61, our model significantly improved performance with a Dice score of 68.40, alongside a reduction in false positive (FPvol: 7.82) and false negative (FNvol: 10.35) volumes.
arXiv Detail & Related papers (2024-09-14T16:39:17Z) - CT-GLIP: 3D Grounded Language-Image Pretraining with CT Scans and Radiology Reports for Full-Body Scenarios [53.94122089629544]
We introduce CT-GLIP (Grounded Language-Image Pretraining with CT scans), a novel method that constructs organ-level image-text pairs to enhance multimodal contrastive learning.
Our method, trained on a multimodal CT dataset comprising 44,011 organ-level vision-text pairs from 17,702 patients across 104 organs, demonstrates it can identify organs and abnormalities in a zero-shot manner using natural languages.
arXiv Detail & Related papers (2024-04-23T17:59:01Z) - 3D Lymphoma Segmentation on PET/CT Images via Multi-Scale Information Fusion with Cross-Attention [6.499725732124126]
This study aims to develop a precise segmentation method for diffuse large B-cell lymphoma (DLBCL) lesions.
We propose a 3D dual-branch encoder segmentation method using shifted window transformers and a Multi-Scale Information Fusion (MSIF) module.
The model was trained and validated on a dataset of 165 DLBCL patients using 5-fold cross-validation.
arXiv Detail & Related papers (2024-02-04T05:25:12Z) - Weakly-Supervised Detection of Bone Lesions in CT [48.34559062736031]
The skeletal region is one of the common sites of metastatic spread of cancer in the breast and prostate.
We developed a pipeline to detect bone lesions in CT volumes via a proxy segmentation task.
Our method detected bone lesions in CT with a precision of 96.7% and recall of 47.3% despite the use of incomplete and partial training data.
arXiv Detail & Related papers (2024-01-31T21:05:34Z) - Whole-body tumor segmentation of 18F -FDG PET/CT using a cascaded and
ensembled convolutional neural networks [2.735686397209314]
The goal of this study was to report the performance of a deep neural network designed to automatically segment regions suspected of cancer in whole-body 18F-FDG PET/CT images.
A cascaded approach was developed where a stacked ensemble of 3D UNET CNN processed the PET/CT images at a fixed 6mm resolution.
arXiv Detail & Related papers (2022-10-14T19:25:56Z) - AutoPET Challenge 2022: Step-by-Step Lesion Segmentation in Whole-body
FDG-PET/CT [0.0]
We propose a novel step-by-step 3D segmentation method to address this problem.
We achieved Dice score of 0.92, false positive volume of 0.89 and false negative volume of 0.53 on preliminary test set.
arXiv Detail & Related papers (2022-09-04T13:49:26Z) - Revisiting 3D Context Modeling with Supervised Pre-training for
Universal Lesion Detection in CT Slices [48.85784310158493]
We propose a Modified Pseudo-3D Feature Pyramid Network (MP3D FPN) to efficiently extract 3D context enhanced 2D features for universal lesion detection in CT slices.
With the novel pre-training method, the proposed MP3D FPN achieves state-of-the-art detection performance on the DeepLesion dataset.
The proposed 3D pre-trained weights can potentially be used to boost the performance of other 3D medical image analysis tasks.
arXiv Detail & Related papers (2020-12-16T07:11:16Z) - M3Lung-Sys: A Deep Learning System for Multi-Class Lung Pneumonia
Screening from CT Imaging [85.00066186644466]
We propose a Multi-task Multi-slice Deep Learning System (M3Lung-Sys) for multi-class lung pneumonia screening from CT imaging.
In addition to distinguish COVID-19 from Healthy, H1N1, and CAP cases, our M 3 Lung-Sys also be able to locate the areas of relevant lesions.
arXiv Detail & Related papers (2020-10-07T06:22:24Z) - Volumetric Attention for 3D Medical Image Segmentation and Detection [53.041572035020344]
A volumetric attention(VA) module for 3D medical image segmentation and detection is proposed.
VA attention is inspired by recent advances in video processing, enables 2.5D networks to leverage context information along the z direction.
Its integration in the Mask R-CNN is shown to enable state-of-the-art performance on the Liver Tumor (LiTS) Challenge.
arXiv Detail & Related papers (2020-04-04T18:55:06Z) - Machine-Learning-Based Multiple Abnormality Prediction with Large-Scale
Chest Computed Tomography Volumes [64.21642241351857]
We curated and analyzed a chest computed tomography (CT) data set of 36,316 volumes from 19,993 unique patients.
We developed a rule-based method for automatically extracting abnormality labels from free-text radiology reports.
We also developed a model for multi-organ, multi-disease classification of chest CT volumes.
arXiv Detail & Related papers (2020-02-12T00:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.