A Contrastive Pretrain Model with Prompt Tuning for Multi-center Medication Recommendation
- URL: http://arxiv.org/abs/2412.20040v1
- Date: Sat, 28 Dec 2024 06:12:02 GMT
- Title: A Contrastive Pretrain Model with Prompt Tuning for Multi-center Medication Recommendation
- Authors: Qidong Liu, Zhaopeng Qiu, Xiangyu Zhao, Xian Wu, Zijian Zhang, Tong Xu, Feng Tian,
- Abstract summary: We introduce a novel conTrastive prEtrain Model with Prompt Tuning (TEMPT) for multi-center medication recommendation.
We devise a novel prompt tuning method to capture the specific information of each hospital rather than adopting the common finetuning.
To validate the proposed model, we conduct extensive experiments on the public eICU multi-center medical dataset.
- Score: 50.43785306804359
- License:
- Abstract: Medication recommendation is one of the most critical health-related applications, which has attracted extensive research interest recently. Most existing works focus on a single hospital with abundant medical data. However, many small hospitals only have a few records, which hinders applying existing medication recommendation works to the real world. Thus, we seek to explore a more practical setting, i.e., multi-center medication recommendation. In this setting, most hospitals have few records, but the total number of records is large. Though small hospitals may benefit from total affluent records, it is also faced with the challenge that the data distributions between various hospitals are much different. In this work, we introduce a novel conTrastive prEtrain Model with Prompt Tuning (TEMPT) for multi-center medication recommendation, which includes two stages of pretraining and finetuning. We first design two self-supervised tasks for the pretraining stage to learn general medical knowledge. They are mask prediction and contrastive tasks, which extract the intra- and inter-relationships of input diagnosis and procedures. Furthermore, we devise a novel prompt tuning method to capture the specific information of each hospital rather than adopting the common finetuning. On the one hand, the proposed prompt tuning can better learn the heterogeneity of each hospital to fit various distributions. On the other hand, it can also relieve the catastrophic forgetting problem of finetuning. To validate the proposed model, we conduct extensive experiments on the public eICU, a multi-center medical dataset. The experimental results illustrate the effectiveness of our model. The implementation code is available to ease the reproducibility https://github.com/Applied-Machine-Learning-Lab/TEMPT.
Related papers
- Continually Evolved Multimodal Foundation Models for Cancer Prognosis [50.43145292874533]
Cancer prognosis is a critical task that involves predicting patient outcomes and survival rates.
Previous studies have integrated diverse data modalities, such as clinical notes, medical images, and genomic data, leveraging their complementary information.
Existing approaches face two major limitations. First, they struggle to incorporate newly arrived data with varying distributions into training, such as patient records from different hospitals.
Second, most multimodal integration methods rely on simplistic concatenation or task-specific pipelines, which fail to capture the complex interdependencies across modalities.
arXiv Detail & Related papers (2025-01-30T06:49:57Z) - Comparing Federated Stochastic Gradient Descent and Federated Averaging for Predicting Hospital Length of Stay [0.0]
Predicting hospital length of stay (LOS) reliably is an essential need for efficient resource allocation at hospitals.
Traditional predictive modeling tools frequently have difficulty acquiring sufficient and diverse data because healthcare institutions have privacy rules in place.
This modeling approach facilitates collaborative model training by modeling decentralized data sources from different hospitals without extracting sensitive data outside of hospitals.
arXiv Detail & Related papers (2024-07-17T17:00:20Z) - NOTE: Notable generation Of patient Text summaries through Efficient
approach based on direct preference optimization [0.0]
"NOTE" stands for "Notable generation Of patient Text summaries through an Efficient approach based on direct preference optimization"
Patient events are sequentially combined and used to generate a discharge summary for each hospitalization.
Note can be utilized to generate various summaries not only discharge summaries but also throughout a patient's journey.
arXiv Detail & Related papers (2024-02-19T06:43:25Z) - SHAPE: A Sample-adaptive Hierarchical Prediction Network for Medication
Recommendation [22.899946140205962]
We propose a novel Sample-adaptive Hierarchical medicAtion Prediction nEtwork, termed SHAPE, to tackle the challenges in the medication recommendation task.
Specifically, we design a compact intra-visit set encoder to encode the relationship in the medical event for obtaining visit-level representation.
To endow the model with the capability of modeling the variable visit length, we introduce a soft curriculum learning method to assign the difficulty of each sample automatically by the visit length.
arXiv Detail & Related papers (2023-09-09T08:28:04Z) - Med-Flamingo: a Multimodal Medical Few-shot Learner [58.85676013818811]
We propose Med-Flamingo, a multimodal few-shot learner adapted to the medical domain.
Based on OpenFlamingo-9B, we continue pre-training on paired and interleaved medical image-text data from publications and textbooks.
We conduct the first human evaluation for generative medical VQA where physicians review the problems and blinded generations in an interactive app.
arXiv Detail & Related papers (2023-07-27T20:36:02Z) - Towards Medical Artificial General Intelligence via Knowledge-Enhanced
Multimodal Pretraining [121.89793208683625]
Medical artificial general intelligence (MAGI) enables one foundation model to solve different medical tasks.
We propose a new paradigm called Medical-knedge-enhanced mulTimOdal pretRaining (MOTOR)
arXiv Detail & Related papers (2023-04-26T01:26:19Z) - Practical Challenges in Differentially-Private Federated Survival
Analysis of Medical Data [57.19441629270029]
In this paper, we take advantage of the inherent properties of neural networks to federate the process of training of survival analysis models.
In the realistic setting of small medical datasets and only a few data centers, this noise makes it harder for the models to converge.
We propose DPFed-post which adds a post-processing stage to the private federated learning scheme.
arXiv Detail & Related papers (2022-02-08T10:03:24Z) - Semi-Supervised Variational Reasoning for Medical Dialogue Generation [70.838542865384]
Two key characteristics are relevant for medical dialogue generation: patient states and physician actions.
We propose an end-to-end variational reasoning approach to medical dialogue generation.
A physician policy network composed of an action-classifier and two reasoning detectors is proposed for augmented reasoning ability.
arXiv Detail & Related papers (2021-05-13T04:14:35Z) - Improving Clinical Outcome Predictions Using Convolution over Medical
Entities with Multimodal Learning [0.522145960878624]
Early prediction of mortality and length of stay(LOS) of a patient is vital for saving a patient's life and management of hospital resources.
In this work, we extract medical entities from clinical notes and use them as additional features besides time-series features to improve our predictions.
We propose a convolution based multimodal architecture, which not only learns effectively combining medical entities and time-series ICU signals of patients.
arXiv Detail & Related papers (2020-11-24T20:08:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.