Optimal path for Biomedical Text Summarization Using Pointer GPT
- URL: http://arxiv.org/abs/2404.08654v1
- Date: Fri, 22 Mar 2024 02:13:23 GMT
- Title: Optimal path for Biomedical Text Summarization Using Pointer GPT
- Authors: Hyunkyung Han, Jaesik Choi,
- Abstract summary: GPT models have a tendency to generate factual errors, lack context, and oversimplify words.
To address these limitations, we replaced the attention mechanism in the GPT model with a pointer network.
The effectiveness of the Pointer-GPT model was evaluated using the ROUGE score.
- Score: 21.919661430250798
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Biomedical text summarization is a critical tool that enables clinicians to effectively ascertain patient status. Traditionally, text summarization has been accomplished with transformer models, which are capable of compressing long documents into brief summaries. However, transformer models are known to be among the most challenging natural language processing (NLP) tasks. Specifically, GPT models have a tendency to generate factual errors, lack context, and oversimplify words. To address these limitations, we replaced the attention mechanism in the GPT model with a pointer network. This modification was designed to preserve the core values of the original text during the summarization process. The effectiveness of the Pointer-GPT model was evaluated using the ROUGE score. The results demonstrated that Pointer-GPT outperformed the original GPT model. These findings suggest that pointer networks can be a valuable addition to EMR systems and can provide clinicians with more accurate and informative summaries of patient medical records. This research has the potential to usher in a new paradigm in EMR systems and to revolutionize the way that clinicians interact with patient medical records.
Related papers
- Improving Expert Radiology Report Summarization by Prompting Large Language Models with a Layperson Summary [8.003346409136348]
Radiology report summarization (RRS) is crucial for patient care, requiring concise "Impressions" from detailed "Findings"
This paper introduces a novel prompting strategy to enhance RRS by first generating a layperson summary.
Our results demonstrate improvements in summarization accuracy and accessibility, particularly in out-of-domain tests.
arXiv Detail & Related papers (2024-06-20T17:01:55Z) - Automated Information Extraction from Thyroid Operation Narrative: A Comparative Study of GPT-4 and Fine-tuned KoELECTRA [1.137357582959183]
This study focuses on the transformative capabilities of the fine-tuned KoELECTRA model in comparison to the GPT-4 model.
The study leverages advanced natural language processing (NLP) techniques to foster a paradigm shift towards more sophisticated data processing systems.
arXiv Detail & Related papers (2024-06-12T06:44:05Z) - Reshaping Free-Text Radiology Notes Into Structured Reports With Generative Transformers [0.29530625605275984]
structured reporting (SR) has been recommended by various medical societies.
We propose a pipeline to extract information from free-text reports.
Our work aims to leverage the potential of Natural Language Processing (NLP) and Transformer-based models.
arXiv Detail & Related papers (2024-03-27T18:38:39Z) - Autocompletion of Chief Complaints in the Electronic Health Records
using Large Language Models [0.3749861135832072]
We utilize text generation techniques to develop machine learning models using Chief Complaint (CC) data.
We tune a prompt by incorporating CC sentences, utilizing the OpenAI API of GPT-4.
We evaluate the models' performance based on the perplexity score, modified BERTScore, and cosine similarity score.
arXiv Detail & Related papers (2024-01-11T18:06:30Z) - PathLDM: Text conditioned Latent Diffusion Model for Histopathology [62.970593674481414]
We introduce PathLDM, the first text-conditioned Latent Diffusion Model tailored for generating high-quality histopathology images.
Our approach fuses image and textual data to enhance the generation process.
We achieved a SoTA FID score of 7.64 for text-to-image generation on the TCGA-BRCA dataset, significantly outperforming the closest text-conditioned competitor with FID 30.1.
arXiv Detail & Related papers (2023-09-01T22:08:32Z) - Customizing General-Purpose Foundation Models for Medical Report
Generation [64.31265734687182]
The scarcity of labelled medical image-report pairs presents great challenges in the development of deep and large-scale neural networks.
We propose customizing off-the-shelf general-purpose large-scale pre-trained models, i.e., foundation models (FMs) in computer vision and natural language processing.
arXiv Detail & Related papers (2023-06-09T03:02:36Z) - Summarizing, Simplifying, and Synthesizing Medical Evidence Using GPT-3
(with Varying Success) [36.646495151276326]
GPT-3 is able to produce high quality summaries of general domain news articles in few- and zero-shot settings.
We enlist domain experts (individuals with medical training) to evaluate summaries of biomedical articles generated by GPT-3, given zero supervision.
arXiv Detail & Related papers (2023-05-10T16:40:37Z) - Does Synthetic Data Generation of LLMs Help Clinical Text Mining? [51.205078179427645]
We investigate the potential of OpenAI's ChatGPT to aid in clinical text mining.
We propose a new training paradigm that involves generating a vast quantity of high-quality synthetic data.
Our method has resulted in significant improvements in the performance of downstream tasks.
arXiv Detail & Related papers (2023-03-08T03:56:31Z) - Cross-modal Clinical Graph Transformer for Ophthalmic Report Generation [116.87918100031153]
We propose a Cross-modal clinical Graph Transformer (CGT) for ophthalmic report generation (ORG)
CGT injects clinical relation triples into the visual features as prior knowledge to drive the decoding procedure.
Experiments on the large-scale FFA-IR benchmark demonstrate that the proposed CGT is able to outperform previous benchmark methods.
arXiv Detail & Related papers (2022-06-04T13:16:30Z) - Towards more patient friendly clinical notes through language models and
ontologies [57.51898902864543]
We present a novel approach to automated medical text based on word simplification and language modelling.
We use a new dataset pairs of publicly available medical sentences and a version of them simplified by clinicians.
Our method based on a language model trained on medical forum data generates simpler sentences while preserving both grammar and the original meaning.
arXiv Detail & Related papers (2021-12-23T16:11:19Z) - An Interpretable End-to-end Fine-tuning Approach for Long Clinical Text [72.62848911347466]
Unstructured clinical text in EHRs contains crucial information for applications including decision support, trial matching, and retrospective research.
Recent work has applied BERT-based models to clinical information extraction and text classification, given these models' state-of-the-art performance in other NLP domains.
In this work, we propose a novel fine-tuning approach called SnipBERT. Instead of using entire notes, SnipBERT identifies crucial snippets and feeds them into a truncated BERT-based model in a hierarchical manner.
arXiv Detail & Related papers (2020-11-12T17:14:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.