NOTE: Notable generation Of patient Text summaries through Efficient
approach based on direct preference optimization
- URL: http://arxiv.org/abs/2402.11882v1
- Date: Mon, 19 Feb 2024 06:43:25 GMT
- Title: NOTE: Notable generation Of patient Text summaries through Efficient
approach based on direct preference optimization
- Authors: Imjin Ahn (1 and 2), Hansle Gwon (1 and 2), Young-Hak Kim (1 and 3),
Tae Joon Jun (1 and 3), Sanghyun Park (2) ((1) INMED DATA, Seoul, Republic of
Korea, (2) Yonsei University, Seoul, Republic of Korea (3) Asan Medical
Center, Seoul, Republic of Korea)
- Abstract summary: "NOTE" stands for "Notable generation Of patient Text summaries through an Efficient approach based on direct preference optimization"
Patient events are sequentially combined and used to generate a discharge summary for each hospitalization.
Note can be utilized to generate various summaries not only discharge summaries but also throughout a patient's journey.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The discharge summary is a one of critical documents in the patient journey,
encompassing all events experienced during hospitalization, including multiple
visits, medications, tests, surgery/procedures, and admissions/discharge.
Providing a summary of the patient's progress is crucial, as it significantly
influences future care and planning. Consequently, clinicians face the
laborious and resource-intensive task of manually collecting, organizing, and
combining all the necessary data for a discharge summary. Therefore, we propose
"NOTE", which stands for "Notable generation Of patient Text summaries through
an Efficient approach based on direct preference optimization". NOTE is based
on Medical Information Mart for Intensive Care- III dataset and summarizes a
single hospitalization of a patient. Patient events are sequentially combined
and used to generate a discharge summary for each hospitalization. In the
present circumstances, large language models' application programming
interfaces (LLMs' APIs) are widely available, but importing and exporting
medical data presents significant challenges due to privacy protection policies
in healthcare institutions. Moreover, to ensure optimal performance, it is
essential to implement a lightweight model for internal server or program
within the hospital. Therefore, we utilized DPO and parameter efficient fine
tuning (PEFT) techniques to apply a fine-tuning method that guarantees superior
performance. To demonstrate the practical application of the developed NOTE, we
provide a webpage-based demonstration software. In the future, we will aim to
deploy the software available for actual use by clinicians in hospital. NOTE
can be utilized to generate various summaries not only discharge summaries but
also throughout a patient's journey, thereby alleviating the labor-intensive
workload of clinicians and aiming for increased efficiency.
Related papers
- STLLaVA-Med: Self-Training Large Language and Vision Assistant for Medical Question-Answering [58.79671189792399]
STLLaVA-Med is designed to train a policy model capable of auto-generating medical visual instruction data.
We validate the efficacy and data efficiency of STLLaVA-Med across three major medical Visual Question Answering (VQA) benchmarks.
arXiv Detail & Related papers (2024-06-28T15:01:23Z) - GAMedX: Generative AI-based Medical Entity Data Extractor Using Large Language Models [1.123722364748134]
This paper introduces GAMedX, a Named Entity Recognition (NER) approach utilizing Large Language Models (LLMs)
The methodology integrates open-source LLMs for NER, utilizing chained prompts and Pydantic schemas for structured output to navigate the complexities of specialized medical jargon.
The findings reveal significant ROUGE F1 score on one of the evaluation datasets with an accuracy of 98%.
arXiv Detail & Related papers (2024-05-31T02:53:22Z) - WisPerMed at "Discharge Me!": Advancing Text Generation in Healthcare with Large Language Models, Dynamic Expert Selection, and Priming Techniques on MIMIC-IV [0.38084074204911494]
This study aims to leverage state of the art language models to automate generating the "Brief Hospital Course" and "Discharge Instructions" sections of Discharge Summaries.
We investigate how automation can improve documentation accuracy, alleviate clinician burnout, and enhance operational efficacy in healthcare facilities.
arXiv Detail & Related papers (2024-05-18T10:56:45Z) - README: Bridging Medical Jargon and Lay Understanding for Patient Education through Data-Centric NLP [9.432205523734707]
We introduce a new task of automatically generating lay definitions, aiming to simplify medical terms into patient-friendly lay language.
We first created the dataset, an extensive collection of over 50,000 unique (medical term, lay definition) pairs and 300,000 mentions.
We have also engineered a data-centric Human-AI pipeline that synergizes data filtering, augmentation, and selection to improve data quality.
arXiv Detail & Related papers (2023-12-24T23:01:00Z) - Clairvoyance: A Pipeline Toolkit for Medical Time Series [95.22483029602921]
Time-series learning is the bread and butter of data-driven *clinical decision support*
Clairvoyance proposes a unified, end-to-end, autoML-friendly pipeline that serves as a software toolkit.
Clairvoyance is the first to demonstrate viability of a comprehensive and automatable pipeline for clinical time-series ML.
arXiv Detail & Related papers (2023-10-28T12:08:03Z) - PULSAR: Pre-training with Extracted Healthcare Terms for Summarising
Patients' Problems and Data Augmentation with Black-box Large Language Models [25.363775123262307]
Automatic summarisation of a patient's problems in the form of a problem list can aid stakeholders in understanding a patient's condition, reducing workload and cognitive bias.
BioNLP 2023 Shared Task 1A focuses on generating a list of diagnoses and problems from the provider's progress notes during hospitalisation.
One component employs large language models (LLMs) for data augmentation; the other is an abstractive summarisation LLM with a novel pre-training objective for generating the patients' problems summarised as a list.
Our approach was ranked second among all submissions to the shared task.
arXiv Detail & Related papers (2023-06-05T10:17:50Z) - SPeC: A Soft Prompt-Based Calibration on Performance Variability of
Large Language Model in Clinical Notes Summarization [50.01382938451978]
We introduce a model-agnostic pipeline that employs soft prompts to diminish variance while preserving the advantages of prompt-based summarization.
Experimental findings indicate that our method not only bolsters performance but also effectively curbs variance for various language models.
arXiv Detail & Related papers (2023-03-23T04:47:46Z) - Optimal discharge of patients from intensive care via a data-driven
policy learning framework [58.720142291102135]
It is important that the patient discharge task addresses the nuanced trade-off between decreasing a patient's length of stay and the risk of readmission or even death following the discharge decision.
This work introduces an end-to-end general framework for capturing this trade-off to recommend optimal discharge timing decisions.
A data-driven approach is used to derive a parsimonious, discrete state space representation that captures a patient's physiological condition.
arXiv Detail & Related papers (2021-12-17T04:39:33Z) - Clinical Outcome Prediction from Admission Notes using Self-Supervised
Knowledge Integration [55.88616573143478]
Outcome prediction from clinical text can prevent doctors from overlooking possible risks.
Diagnoses at discharge, procedures performed, in-hospital mortality and length-of-stay prediction are four common outcome prediction targets.
We propose clinical outcome pre-training to integrate knowledge about patient outcomes from multiple public sources.
arXiv Detail & Related papers (2021-02-08T10:26:44Z) - BiteNet: Bidirectional Temporal Encoder Network to Predict Medical
Outcomes [53.163089893876645]
We propose a novel self-attention mechanism that captures the contextual dependency and temporal relationships within a patient's healthcare journey.
An end-to-end bidirectional temporal encoder network (BiteNet) then learns representations of the patient's journeys.
We have evaluated the effectiveness of our methods on two supervised prediction and two unsupervised clustering tasks with a real-world EHR dataset.
arXiv Detail & Related papers (2020-09-24T00:42:36Z) - Self-Attention Enhanced Patient Journey Understanding in Healthcare
System [43.11457142941327]
MusaNet is designed to learn the representations of patient journeys that is used to be a long sequence of activities.
The MusaNet is trained in end-to-end manner using the training data derived from EHRs.
Results have demonstrated the proposed MusaNet produces higher-quality representations than state-of-the-art baseline methods.
arXiv Detail & Related papers (2020-06-15T10:32:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.