Curriculum Based Multi-Task Learning for Parkinson's Disease Detection
- URL: http://arxiv.org/abs/2302.13631v1
- Date: Mon, 27 Feb 2023 09:58:09 GMT
- Title: Curriculum Based Multi-Task Learning for Parkinson's Disease Detection
- Authors: Nikhil J. Dhinagar, Conor Owens-Walton, Emily Laltoo, Christina P.
Boyle, Yao-Liang Chen, Philip Cook, Corey McMillan, Chih-Chien Tsai, J-J
Wang, Yih-Ru Wu, Ysbrand van der Werf, Paul M. Thompson
- Abstract summary: We leverage severity-based meta-data on the stages of disease to define a curriculum for training a deep convolutional neural network (CNN)
curriculum learning is a training strategy that aims to boost classifier performance by starting with examples that are easier to classify.
Even with our multi-task setting using pre-trained CNNs and transfer learning, PD classification based on T1-weighted (T1-w) MRI was challenging (ROC AUC: 0.59-0.65), but curriculum training boosted performance (by 3.9%) compared to our baseline model.
- Score: 2.1187904593676845
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There is great interest in developing radiological classifiers for diagnosis,
staging, and predictive modeling in progressive diseases such as Parkinson's
disease (PD), a neurodegenerative disease that is difficult to detect in its
early stages. Here we leverage severity-based meta-data on the stages of
disease to define a curriculum for training a deep convolutional neural network
(CNN). Typically, deep learning networks are trained by randomly selecting
samples in each mini-batch. By contrast, curriculum learning is a training
strategy that aims to boost classifier performance by starting with examples
that are easier to classify. Here we define a curriculum to progressively
increase the difficulty of the training data corresponding to the Hoehn and
Yahr (H&Y) staging system for PD (total N=1,012; 653 PD patients, 359 controls;
age range: 20.0-84.9 years). Even with our multi-task setting using pre-trained
CNNs and transfer learning, PD classification based on T1-weighted (T1-w) MRI
was challenging (ROC AUC: 0.59-0.65), but curriculum training boosted
performance (by 3.9%) compared to our baseline model. Future work with
multimodal imaging may further boost performance.
Related papers
- Brain Tumor Classification on MRI in Light of Molecular Markers [61.77272414423481]
Co-deletion of the 1p/19q gene is associated with clinical outcomes in low-grade gliomas.
This study aims to utilize a specially MRI-based convolutional neural network for brain cancer detection.
arXiv Detail & Related papers (2024-09-29T07:04:26Z) - Applications of Sequential Learning for Medical Image Classification [0.13191970195165517]
We develop a neural network training framework for continual training of small amounts of medical imaging data.
We address problems that impede sequential learning such as overfitting, catastrophic forgetting, and concept drift.
arXiv Detail & Related papers (2023-09-26T00:46:25Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Video and Synthetic MRI Pre-training of 3D Vision Architectures for
Neuroimage Analysis [3.208731414009847]
Transfer learning involves pre-training deep learning models on a large corpus of data for adaptation to specific tasks.
We benchmarked vision transformers (ViTs) and convolutional neural networks (CNNs) with varied upstream pre-training approaches.
The resulting pre-trained models can be adapted to a range of downstream tasks, even when training data for the target task is limited.
arXiv Detail & Related papers (2023-09-09T00:33:23Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Continual Learning with Bayesian Model based on a Fixed Pre-trained
Feature Extractor [55.9023096444383]
Current deep learning models are characterised by catastrophic forgetting of old knowledge when learning new classes.
Inspired by the process of learning new knowledge in human brains, we propose a Bayesian generative model for continual learning.
arXiv Detail & Related papers (2022-04-28T08:41:51Z) - Multi-task fusion for improving mammography screening data
classification [3.7683182861690843]
We propose a pipeline approach, where we first train a set of individual, task-specific models.
We then investigate the fusion thereof, which is in contrast to the standard model ensembling strategy.
Our fusion approaches improve AUC scores significantly by up to 0.04 compared to standard model ensembling.
arXiv Detail & Related papers (2021-12-01T13:56:27Z) - Pre-training transformer-based framework on large-scale pediatric claims
data for downstream population-specific tasks [3.1580072841682734]
This study presents the Claim Pre-Training (Claim-PT) framework, a generic pre-training model that first trains on the entire pediatric claims dataset.
The effective knowledge transfer is completed through the task-aware fine-tuning stage.
We conducted experiments on a real-world claims dataset with more than one million patient records.
arXiv Detail & Related papers (2021-06-24T15:25:41Z) - Transfer Learning with Ensembles of Deep Neural Networks for Skin Cancer
Classification in Imbalanced Data Sets [0.6802401545890961]
Machine learning techniques for accurate classification of skin cancer from medical images have been reported.
Many techniques are based on pre-trained convolutional neural networks (CNNs), which enable training the models based on limited amounts of training data.
We propose a novel ensemble-based CNN architecture where multiple CNN models, some of which are pre-trained and some are trained only on the data at hand, along with patient information (meta-data) are combined using a meta-learner.
arXiv Detail & Related papers (2021-03-22T06:04:45Z) - Classification of COVID-19 in CT Scans using Multi-Source Transfer
Learning [91.3755431537592]
We propose the use of Multi-Source Transfer Learning to improve upon traditional Transfer Learning for the classification of COVID-19 from CT scans.
With our multi-source fine-tuning approach, our models outperformed baseline models fine-tuned with ImageNet.
Our best performing model was able to achieve an accuracy of 0.893 and a Recall score of 0.897, outperforming its baseline Recall score by 9.3%.
arXiv Detail & Related papers (2020-09-22T11:53:06Z) - Select-ProtoNet: Learning to Select for Few-Shot Disease Subtype
Prediction [55.94378672172967]
We focus on few-shot disease subtype prediction problem, identifying subgroups of similar patients.
We introduce meta learning techniques to develop a new model, which can extract the common experience or knowledge from interrelated clinical tasks.
Our new model is built upon a carefully designed meta-learner, called Prototypical Network, that is a simple yet effective meta learning machine for few-shot image classification.
arXiv Detail & Related papers (2020-09-02T02:50:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.