Adversarial Multi-Source Transfer Learning in Healthcare: Application to
Glucose Prediction for Diabetic People
- URL: http://arxiv.org/abs/2006.15940v1
- Date: Mon, 29 Jun 2020 11:17:50 GMT
- Title: Adversarial Multi-Source Transfer Learning in Healthcare: Application to
Glucose Prediction for Diabetic People
- Authors: Maxime De Bois, Moun\^im A. El Yacoubi, and Mehdi Ammi
- Abstract summary: We propose a multi-source adversarial transfer learning framework that enables the learning of a feature representation that is similar across the sources.
We apply this idea to glucose forecasting for diabetic people using a fully convolutional neural network.
In particular, it shines when using data from different datasets, or when there is too little data in an intra-dataset situation.
- Score: 4.17510581764131
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning has yet to revolutionize general practices in healthcare,
despite promising results for some specific tasks. This is partly due to data
being in insufficient quantities hurting the training of the models. To address
this issue, data from multiple health actors or patients could be combined by
capitalizing on their heterogeneity through the use of transfer learning.
To improve the quality of the transfer between multiple sources of data, we
propose a multi-source adversarial transfer learning framework that enables the
learning of a feature representation that is similar across the sources, and
thus more general and more easily transferable. We apply this idea to glucose
forecasting for diabetic people using a fully convolutional neural network. The
evaluation is done by exploring various transfer scenarios with three datasets
characterized by their high inter and intra variability.
While transferring knowledge is beneficial in general, we show that the
statistical and clinical accuracies can be further improved by using of the
adversarial training methodology, surpassing the current state-of-the-art
results. In particular, it shines when using data from different datasets, or
when there is too little data in an intra-dataset situation. To understand the
behavior of the models, we analyze the learnt feature representations and
propose a new metric in this regard. Contrary to a standard transfer, the
adversarial transfer does not discriminate the patients and datasets, helping
the learning of a more general feature representation.
The adversarial training framework improves the learning of a general feature
representation in a multi-source environment, enhancing the knowledge transfer
to an unseen target.
The proposed method can help improve the efficiency of data shared by
different health actors in the training of deep models.
Related papers
- An Experimental Comparison of Transfer Learning against Self-supervised Learning [6.744847405966574]
This paper compares the performance and robustness of transfer learning and self-supervised learning in the medical field.
We tested data with several common issues in medical domains, such as data imbalance, data scarcity, and domain mismatch.
We provide recommendations to help users apply transfer learning and self-supervised learning methods in medical areas.
arXiv Detail & Related papers (2024-07-08T04:14:52Z) - Transfer learning from a sparsely annotated dataset of 3D medical images [4.477071833136902]
This study explores the use of transfer learning to improve the performance of deep convolutional neural networks for organ segmentation in medical imaging.
A base segmentation model was trained on a large and sparsely annotated dataset; its weights were used for transfer learning on four new down-stream segmentation tasks.
The results showed that transfer learning from the base model was beneficial when small datasets were available.
arXiv Detail & Related papers (2023-11-08T21:31:02Z) - Amplifying Pathological Detection in EEG Signaling Pathways through
Cross-Dataset Transfer Learning [10.212217551908525]
We study the effectiveness of data and model scaling and cross-dataset knowledge transfer in a real-world pathology classification task.
We identify the challenges of possible negative transfer and emphasize the significance of some key components.
Our findings indicate a small and generic model (e.g. ShallowNet) performs well on a single dataset, however, a larger model (e.g. TCN) performs better on transfer and learning from a larger and diverse dataset.
arXiv Detail & Related papers (2023-09-19T20:09:15Z) - Incomplete Multimodal Learning for Complex Brain Disorders Prediction [65.95783479249745]
We propose a new incomplete multimodal data integration approach that employs transformers and generative adversarial networks.
We apply our new method to predict cognitive degeneration and disease outcomes using the multimodal imaging genetic data from Alzheimer's Disease Neuroimaging Initiative cohort.
arXiv Detail & Related papers (2023-05-25T16:29:16Z) - Knowledge-Guided Data-Centric AI in Healthcare: Progress, Shortcomings,
and Future Directions [7.673853485227739]
Deep learning is able to learn a wide range of examples of a particular concept or meaning.
In medicine, having a diverse set of training data on a particular disease can lead to the development of a model that is able to accurately predict the disease.
Despite the potential benefits, there have not been significant advances in image-based diagnosis due to a lack of high-quality annotated data.
This article highlights the importance of using a data-centric approach to improve the quality of data representations.
arXiv Detail & Related papers (2022-12-27T19:35:14Z) - When Accuracy Meets Privacy: Two-Stage Federated Transfer Learning
Framework in Classification of Medical Images on Limited Data: A COVID-19
Case Study [77.34726150561087]
COVID-19 pandemic has spread rapidly and caused a shortage of global medical resources.
CNN has been widely utilized and verified in analyzing medical images.
arXiv Detail & Related papers (2022-03-24T02:09:41Z) - BERT WEAVER: Using WEight AVERaging to enable lifelong learning for
transformer-based models in biomedical semantic search engines [49.75878234192369]
We present WEAVER, a simple, yet efficient post-processing method that infuses old knowledge into the new model.
We show that applying WEAVER in a sequential manner results in similar word embedding distributions as doing a combined training on all data at once.
arXiv Detail & Related papers (2022-02-21T10:34:41Z) - Practical Challenges in Differentially-Private Federated Survival
Analysis of Medical Data [57.19441629270029]
In this paper, we take advantage of the inherent properties of neural networks to federate the process of training of survival analysis models.
In the realistic setting of small medical datasets and only a few data centers, this noise makes it harder for the models to converge.
We propose DPFed-post which adds a post-processing stage to the private federated learning scheme.
arXiv Detail & Related papers (2022-02-08T10:03:24Z) - Domain Generalization for Medical Imaging Classification with
Linear-Dependency Regularization [59.5104563755095]
We introduce a simple but effective approach to improve the generalization capability of deep neural networks in the field of medical imaging classification.
Motivated by the observation that the domain variability of the medical images is to some extent compact, we propose to learn a representative feature space through variational encoding.
arXiv Detail & Related papers (2020-09-27T12:30:30Z) - Select-ProtoNet: Learning to Select for Few-Shot Disease Subtype
Prediction [55.94378672172967]
We focus on few-shot disease subtype prediction problem, identifying subgroups of similar patients.
We introduce meta learning techniques to develop a new model, which can extract the common experience or knowledge from interrelated clinical tasks.
Our new model is built upon a carefully designed meta-learner, called Prototypical Network, that is a simple yet effective meta learning machine for few-shot image classification.
arXiv Detail & Related papers (2020-09-02T02:50:30Z) - Unlabeled Data Deployment for Classification of Diabetic Retinopathy
Images Using Knowledge Transfer [11.031841470875571]
Transfer learning is used to solve the problem of lack of labeled data.
Knowledge distillation is recently proposed to transfer the knowledge of a model to another one.
In this paper, a novel knowledge distillation using transfer learning is proposed to transfer the whole knowledge of a model to another one.
arXiv Detail & Related papers (2020-02-09T09:01:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.