Towards General Purpose Medical AI: Continual Learning Medical
Foundation Model
- URL: http://arxiv.org/abs/2303.06580v1
- Date: Sun, 12 Mar 2023 05:27:22 GMT
- Title: Towards General Purpose Medical AI: Continual Learning Medical
Foundation Model
- Authors: Huahui Yi, Ziyuan Qin, Qicheng Lao, Wei Xu, Zekun Jiang, Dequan Wang,
Shaoting Zhang, Kang Li
- Abstract summary: Inevitable domain and task discrepancies in real-world scenarios can impair the generalization performance of the pre-trained deep models for medical data.
We audaciously propose that we should build a general-purpose medical AI system that can be seamlessly adapted to downstream domains/tasks.
- Score: 22.03086588403621
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Inevitable domain and task discrepancies in real-world scenarios can impair
the generalization performance of the pre-trained deep models for medical data.
Therefore, we audaciously propose that we should build a general-purpose
medical AI system that can be seamlessly adapted to downstream domains/tasks.
Since the domain/task adaption procedures usually involve additional labeling
work for the target data, designing a data-efficient adaption algorithm is
desired to save the cost of transferring the learned knowledge. Our recent work
found that vision-language models (VLMs) are efficient learners with
extraordinary cross-domain ability. Therefore, in this work, we further explore
the possibility of leveraging pre-trained VLMs as medical foundation models for
building general-purpose medical AI, where we thoroughly investigate three
machine-learning paradigms, i.e., domain/task-specialized learning, joint
learning, and continual learning, for training the VLMs and evaluate their
generalization performance on cross-domain and cross-task test sets. To
alleviate the catastrophic forgetting during sequential training, we employ
rehearsal learning and receive a sharp boost in terms of generalization
capability. In a nutshell, our empirical evidence suggests that continual
learning may be a practical and efficient learning paradigm for the medical
foundation model. And we hope researchers can use our empirical evidence as
basement to further explore the path toward medical foundation model.
Related papers
- An Efficient Contrastive Unimodal Pretraining Method for EHR Time Series Data [35.943089444017666]
We propose an efficient method of contrastive pretraining tailored for long clinical timeseries data.
Our model demonstrates the ability to impute missing measurements, providing clinicians with deeper insights into patient conditions.
arXiv Detail & Related papers (2024-10-11T19:05:25Z) - Towards a clinically accessible radiology foundation model: open-access and lightweight, with automated evaluation [113.5002649181103]
Training open-source small multimodal models (SMMs) to bridge competency gaps for unmet clinical needs in radiology.
For training, we assemble a large dataset of over 697 thousand radiology image-text pairs.
For evaluation, we propose CheXprompt, a GPT-4-based metric for factuality evaluation, and demonstrate its parity with expert evaluation.
The inference of LlaVA-Rad is fast and can be performed on a single V100 GPU in private settings, offering a promising state-of-the-art tool for real-world clinical applications.
arXiv Detail & Related papers (2024-03-12T18:12:02Z) - ProtoKD: Learning from Extremely Scarce Data for Parasite Ova
Recognition [5.224806515926022]
We introduce ProtoKD, one of the first approaches to tackle the problem of multi-class parasitic ova recognition using extremely scarce data.
We establish a new benchmark to drive research in this critical direction and validate that the proposed ProtoKD framework achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-09-18T23:49:04Z) - Incomplete Multimodal Learning for Complex Brain Disorders Prediction [65.95783479249745]
We propose a new incomplete multimodal data integration approach that employs transformers and generative adversarial networks.
We apply our new method to predict cognitive degeneration and disease outcomes using the multimodal imaging genetic data from Alzheimer's Disease Neuroimaging Initiative cohort.
arXiv Detail & Related papers (2023-05-25T16:29:16Z) - Towards Medical Artificial General Intelligence via Knowledge-Enhanced
Multimodal Pretraining [121.89793208683625]
Medical artificial general intelligence (MAGI) enables one foundation model to solve different medical tasks.
We propose a new paradigm called Medical-knedge-enhanced mulTimOdal pretRaining (MOTOR)
arXiv Detail & Related papers (2023-04-26T01:26:19Z) - Federated Cycling (FedCy): Semi-supervised Federated Learning of
Surgical Phases [57.90226879210227]
FedCy is a semi-supervised learning (FSSL) method that combines FL and self-supervised learning to exploit a decentralized dataset of both labeled and unlabeled videos.
We demonstrate significant performance gains over state-of-the-art FSSL methods on the task of automatic recognition of surgical phases.
arXiv Detail & Related papers (2022-03-14T17:44:53Z) - BERT WEAVER: Using WEight AVERaging to enable lifelong learning for
transformer-based models in biomedical semantic search engines [49.75878234192369]
We present WEAVER, a simple, yet efficient post-processing method that infuses old knowledge into the new model.
We show that applying WEAVER in a sequential manner results in similar word embedding distributions as doing a combined training on all data at once.
arXiv Detail & Related papers (2022-02-21T10:34:41Z) - Pre-training transformer-based framework on large-scale pediatric claims
data for downstream population-specific tasks [3.1580072841682734]
This study presents the Claim Pre-Training (Claim-PT) framework, a generic pre-training model that first trains on the entire pediatric claims dataset.
The effective knowledge transfer is completed through the task-aware fine-tuning stage.
We conducted experiments on a real-world claims dataset with more than one million patient records.
arXiv Detail & Related papers (2021-06-24T15:25:41Z) - Domain Generalization on Medical Imaging Classification using Episodic
Training with Task Augmentation [62.49837463676111]
We propose a novel scheme of episodic training with task augmentation on medical imaging classification.
Motivated by the limited number of source domains in real-world medical deployment, we consider the unique task-level overfitting.
arXiv Detail & Related papers (2021-06-13T03:56:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.