Effect of Intensity Standardization on Deep Learning for WML
Segmentation in Multi-Centre FLAIR MRI
- URL: http://arxiv.org/abs/2307.03827v1
- Date: Fri, 7 Jul 2023 20:51:38 GMT
- Title: Effect of Intensity Standardization on Deep Learning for WML
Segmentation in Multi-Centre FLAIR MRI
- Authors: Abdollah Ghazvanchahi, Pejman Jahbedar Maralani, Alan R. Moody, April
Khademi
- Abstract summary: Deep learning (DL) methods for white matter lesion (WML) segmentation in MRI suffer a reduction in performance when applied on data from a scanner or centre that is out-of-distribution (OOD) from the training data.
This is critical for translation and widescale adoption, since current models cannot be readily applied to data from new institutions.
We evaluate several intensity standardization methods for MRI as a preprocessing step for WML segmentation in multi-centre Fluid-Attenuated Inversion Recovery (FLAIR) MRI.
- Score: 0.06117371161379209
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning (DL) methods for white matter lesion (WML) segmentation in MRI
suffer a reduction in performance when applied on data from a scanner or centre
that is out-of-distribution (OOD) from the training data. This is critical for
translation and widescale adoption, since current models cannot be readily
applied to data from new institutions. In this work, we evaluate several
intensity standardization methods for MRI as a preprocessing step for WML
segmentation in multi-centre Fluid-Attenuated Inversion Recovery (FLAIR) MRI.
We evaluate a method specifically developed for FLAIR MRI called IAMLAB along
with other popular normalization techniques such as White-strip, Nyul and
Z-score. We proposed an Ensemble model that combines predictions from each of
these models. A skip-connection UNet (SC UNet) was trained on the standardized
images, as well as the original data and segmentation performance was evaluated
over several dimensions. The training (in-distribution) data consists of a
single study, of 60 volumes, and the test (OOD) data is 128 unseen volumes from
three clinical cohorts. Results show IAMLAB and Ensemble provide higher WML
segmentation performance compared to models from original data or other
normalization methods. IAMLAB & Ensemble have the highest dice similarity
coefficient (DSC) on the in-distribution data (0.78 & 0.80) and on clinical OOD
data. DSC was significantly higher for IAMLAB compared to the original data
(p<0.05) for all lesion categories (LL>25mL: 0.77 vs. 0.71; 10mL<= LL<25mL:
0.66 vs. 0.61; LL<10mL: 0.53 vs. 0.52). The IAMLAB and Ensemble normalization
methods are mitigating MRI domain shift and are optimal for DL-based WML
segmentation in unseen FLAIR data.
Related papers
- Entropy Law: The Story Behind Data Compression and LLM Performance [115.70395740286422]
We find that model performance is negatively correlated to the compression ratio of training data, which usually yields a lower training loss.
Based on the findings of the entropy law, we propose a quite efficient and universal data selection method.
We also present an interesting application of entropy law that can detect potential performance risks at the beginning of model training.
arXiv Detail & Related papers (2024-07-09T08:14:29Z) - DataComp-LM: In search of the next generation of training sets for language models [200.5293181577585]
DataComp for Language Models (DCLM) is a testbed for controlled dataset experiments with the goal of improving language models.
We provide a standardized corpus of 240T tokens extracted from Common Crawl, effective pretraining recipes based on the OpenLM framework, and a broad suite of 53 downstream evaluations.
Participants in the DCLM benchmark can experiment with data curation strategies such as deduplication, filtering, and data mixing at model scales ranging from 412M to 7B parameters.
arXiv Detail & Related papers (2024-06-17T17:42:57Z) - Multicenter Privacy-Preserving Model Training for Deep Learning Brain Metastases Autosegmentation [2.479757014250359]
This work aims to explore the impact of multicenter data heterogeneity on deep learning brain metastases (BM) autosegmentation performance.
incremental transfer learning technique, namely learning without forgetting (LWF), to improve model generalizability without sharing raw data.
When the UKER pretrained model is applied to USZ, LWF achieves a higher average F1 score (0.839) than naive TL (0.570) and single-center training (0.688) on combined UKER and USZ test data.
arXiv Detail & Related papers (2024-05-17T16:01:11Z) - Quantifying uncertainty in lung cancer segmentation with foundation models applied to mixed-domain datasets [6.712251433139412]
Medical image foundation models have shown the ability to segment organs and tumors with minimal fine-tuning.
These models are typically evaluated on task-specific in-distribution (ID) datasets.
We introduced a comprehensive set of computationally fast metrics to evaluate the performance of multiple foundation models trained with self-supervised learning (SSL)
SMIT produced a highest F1-score (LRAD: 0.60, 5Rater: 0.64) and lowest entropy (LRAD: 0.06, 5Rater: 0.12), indicating higher tumor detection rate and confident segmentations.
arXiv Detail & Related papers (2024-03-19T19:36:48Z) - Better Generalization of White Matter Tract Segmentation to Arbitrary
Datasets with Scaled Residual Bootstrap [1.30536490219656]
White matter (WM) tract segmentation is a crucial step for brain connectivity studies.
We propose a WM tract segmentation approach that improves the generalization with scaled residual bootstrap.
arXiv Detail & Related papers (2023-09-25T09:31:34Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Hybrid Representation-Enhanced Sampling for Bayesian Active Learning in
Musculoskeletal Segmentation of Lower Extremities [0.9287179270753105]
This study introduces a hybrid representation-enhanced sampling strategy that integrates both density and diversity criteria.
Experiments are performed on two lower extremity (LE) datasets of MRI and CT images.
arXiv Detail & Related papers (2023-07-26T06:52:29Z) - Learnable Weight Initialization for Volumetric Medical Image Segmentation [66.3030435676252]
We propose a learnable weight-based hybrid medical image segmentation approach.
Our approach is easy to integrate into any hybrid model and requires no external training data.
Experiments on multi-organ and lung cancer segmentation tasks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-06-15T17:55:05Z) - MAPPING: Model Average with Post-processing for Stroke Lesion
Segmentation [57.336056469276585]
We present our stroke lesion segmentation model based on nnU-Net framework, and apply it to the Anatomical Tracings of Lesions After Stroke dataset.
Our method took the first place in the 2022 MICCAI ATLAS Challenge with an average Dice score of 0.6667, Lesion-wise F1 score of 0.5643, Simple Lesion Count score of 4.5367, and Volume Difference score of 8804.9102.
arXiv Detail & Related papers (2022-11-11T14:17:04Z) - A Deep Learning-based Method to Extract Lumen and Media-Adventitia in
Intravascular Ultrasound Images [3.2963079183841297]
Intravascular ultrasound (IVUS) imaging allows direct visualization of the coronary vessel wall.
Current segmentation relies on manual operations, which is time-consuming and user-dependent.
In this paper, we aim to develop a deep learning-based method using an encoder-decoder deep architecture.
arXiv Detail & Related papers (2021-02-21T00:10:05Z) - Federated Deep AUC Maximization for Heterogeneous Data with a Constant
Communication Complexity [77.78624443410216]
We propose improved FDAM algorithms for detecting heterogeneous chest data.
A result of this paper is that the communication of the proposed algorithm is strongly independent of the number of machines and also independent of the accuracy level.
Experiments have demonstrated the effectiveness of our FDAM algorithm on benchmark datasets and on medical chest Xray images from different organizations.
arXiv Detail & Related papers (2021-02-09T04:05:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.