Transfer learning from a sparsely annotated dataset of 3D medical images
- URL: http://arxiv.org/abs/2311.05032v1
- Date: Wed, 8 Nov 2023 21:31:02 GMT
- Title: Transfer learning from a sparsely annotated dataset of 3D medical images
- Authors: Gabriel Efrain Humpire-Mamani, Colin Jacobs, Mathias Prokop, Bram van
Ginneken, Nikolas Lessmann
- Abstract summary: This study explores the use of transfer learning to improve the performance of deep convolutional neural networks for organ segmentation in medical imaging.
A base segmentation model was trained on a large and sparsely annotated dataset; its weights were used for transfer learning on four new down-stream segmentation tasks.
The results showed that transfer learning from the base model was beneficial when small datasets were available.
- Score: 4.477071833136902
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Transfer learning leverages pre-trained model features from a large dataset
to save time and resources when training new models for various tasks,
potentially enhancing performance. Due to the lack of large datasets in the
medical imaging domain, transfer learning from one medical imaging model to
other medical imaging models has not been widely explored. This study explores
the use of transfer learning to improve the performance of deep convolutional
neural networks for organ segmentation in medical imaging. A base segmentation
model (3D U-Net) was trained on a large and sparsely annotated dataset; its
weights were used for transfer learning on four new down-stream segmentation
tasks for which a fully annotated dataset was available. We analyzed the
training set size's influence to simulate scarce data. The results showed that
transfer learning from the base model was beneficial when small datasets were
available, providing significant performance improvements; where fine-tuning
the base model is more beneficial than updating all the network weights with
vanilla transfer learning. Transfer learning with fine-tuning increased the
performance by up to 0.129 (+28\%) Dice score than experiments trained from
scratch, and on average 23 experiments increased the performance by 0.029 Dice
score in the new segmentation tasks. The study also showed that cross-modality
transfer learning using CT scans was beneficial. The findings of this study
demonstrate the potential of transfer learning to improve the efficiency of
annotation and increase the accessibility of accurate organ segmentation in
medical imaging, ultimately leading to improved patient care. We made the
network definition and weights publicly available to benefit other users and
researchers.
Related papers
- Computational Pathology at Health System Scale -- Self-Supervised
Foundation Models from Three Billion Images [30.618749295623363]
This project aims to train the largest academic foundation model and benchmark the most prominent self-supervised learning algorithms by pre-training.
We collected the largest pathology dataset to date, consisting of over 3 billion images from over 423 thousand microscopy slides.
Our results demonstrate that pre-training on pathology data is beneficial for downstream performance compared to pre-training on natural images.
arXiv Detail & Related papers (2023-10-10T21:40:19Z) - Understanding the Tricks of Deep Learning in Medical Image Segmentation:
Challenges and Future Directions [66.40971096248946]
In this paper, we collect a series of MedISeg tricks for different model implementation phases.
We experimentally explore the effectiveness of these tricks on consistent baselines.
We also open-sourced a strong MedISeg repository, where each component has the advantage of plug-and-play.
arXiv Detail & Related papers (2022-09-21T12:30:05Z) - BERT WEAVER: Using WEight AVERaging to enable lifelong learning for
transformer-based models in biomedical semantic search engines [49.75878234192369]
We present WEAVER, a simple, yet efficient post-processing method that infuses old knowledge into the new model.
We show that applying WEAVER in a sequential manner results in similar word embedding distributions as doing a combined training on all data at once.
arXiv Detail & Related papers (2022-02-21T10:34:41Z) - Medulloblastoma Tumor Classification using Deep Transfer Learning with
Multi-Scale EfficientNets [63.62764375279861]
We propose an end-to-end MB tumor classification and explore transfer learning with various input sizes and matching network dimensions.
Using a data set with 161 cases, we demonstrate that pre-trained EfficientNets with larger input resolutions lead to significant performance improvements.
arXiv Detail & Related papers (2021-09-10T13:07:11Z) - On the Robustness of Pretraining and Self-Supervision for a Deep
Learning-based Analysis of Diabetic Retinopathy [70.71457102672545]
We compare the impact of different training procedures for diabetic retinopathy grading.
We investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions.
Our results indicate that models from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions.
arXiv Detail & Related papers (2021-06-25T08:32:45Z) - Transfer Learning Enhanced Generative Adversarial Networks for
Multi-Channel MRI Reconstruction [3.5765797841178597]
Deep learning based generative adversarial networks (GAN) can effectively perform image reconstruction with under-sampled MR data.
It is difficult to obtain tens of thousands of raw patient data to train the model since saving k-space data is not in the routine clinical flow.
In this study, three novel applications were explored based on parallel imaging combined with the GAN model (PI-GAN) and transfer learning.
arXiv Detail & Related papers (2021-05-17T21:28:00Z) - A Multi-Stage Attentive Transfer Learning Framework for Improving
COVID-19 Diagnosis [49.3704402041314]
We propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis.
Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains.
Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images.
arXiv Detail & Related papers (2021-01-14T01:39:19Z) - Self supervised contrastive learning for digital histopathology [0.0]
We use a contrastive self-supervised learning method called SimCLR that achieved state-of-the-art results on natural-scene images.
We find that combining multiple multi-organ datasets with different types of staining and resolution properties improves the quality of the learned features.
Linear classifiers trained on top of the learned features show that networks pretrained on digital histopathology datasets perform better than ImageNet pretrained networks.
arXiv Detail & Related papers (2020-11-27T19:18:45Z) - Bidirectional RNN-based Few Shot Learning for 3D Medical Image
Segmentation [11.873435088539459]
We propose a 3D few shot segmentation framework for accurate organ segmentation using limited training samples of the target organ annotation.
A U-Net like network is designed to predict segmentation by learning the relationship between 2D slices of support data and a query image.
We evaluate our proposed model using three 3D CT datasets with annotations of different organs.
arXiv Detail & Related papers (2020-11-19T01:44:55Z) - Adversarial Multi-Source Transfer Learning in Healthcare: Application to
Glucose Prediction for Diabetic People [4.17510581764131]
We propose a multi-source adversarial transfer learning framework that enables the learning of a feature representation that is similar across the sources.
We apply this idea to glucose forecasting for diabetic people using a fully convolutional neural network.
In particular, it shines when using data from different datasets, or when there is too little data in an intra-dataset situation.
arXiv Detail & Related papers (2020-06-29T11:17:50Z) - 3D medical image segmentation with labeled and unlabeled data using
autoencoders at the example of liver segmentation in CT images [58.720142291102135]
This work investigates the potential of autoencoder-extracted features to improve segmentation with a convolutional neural network.
A convolutional autoencoder was used to extract features from unlabeled data and a multi-scale, fully convolutional CNN was used to perform the target task of 3D liver segmentation in CT images.
arXiv Detail & Related papers (2020-03-17T20:20:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.