ChatGPT Makes Medicine Easy to Swallow: An Exploratory Case Study on
Simplified Radiology Reports
- URL: http://arxiv.org/abs/2212.14882v1
- Date: Fri, 30 Dec 2022 18:55:16 GMT
- Title: ChatGPT Makes Medicine Easy to Swallow: An Exploratory Case Study on
Simplified Radiology Reports
- Authors: Katharina Jeblick, Balthasar Schachtner, Jakob Dexl, Andreas
Mittermeier, Anna Theresa St\"uber, Johanna Topalis, Tobias Weber, Philipp
Wesp, Bastian Sabel, Jens Ricke, Michael Ingrisch
- Abstract summary: ChatGPT is a language model capable of generating text that appears human-like and authentic.
We asked 15 radiologists to assess the quality of radiology reports simplified by ChatGPT.
Most radiologists agreed that the simplified reports were factually correct, complete, and not potentially harmful to the patient.
- Score: 0.4194454151396506
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The release of ChatGPT, a language model capable of generating text that
appears human-like and authentic, has gained significant attention beyond the
research community. We expect that the convincing performance of ChatGPT
incentivizes users to apply it to a variety of downstream tasks, including
prompting the model to simplify their own medical reports. To investigate this
phenomenon, we conducted an exploratory case study. In a questionnaire, we
asked 15 radiologists to assess the quality of radiology reports simplified by
ChatGPT. Most radiologists agreed that the simplified reports were factually
correct, complete, and not potentially harmful to the patient. Nevertheless,
instances of incorrect statements, missed key medical findings, and potentially
harmful passages were reported. While further studies are needed, the initial
insights of this study indicate a great potential in using large language
models like ChatGPT to improve patient-centered care in radiology and other
medical domains.
Related papers
- ReXplain: Translating Radiology into Patient-Friendly Video Reports [5.787653511498558]
ReXplain is an AI-driven system that generates patient-friendly video reports for radiology findings.
Our proof-of-concept study with five board-certified radiologists indicates that ReXplain could accurately deliver radiological information.
arXiv Detail & Related papers (2024-10-01T06:41:18Z) - Two-Pronged Human Evaluation of ChatGPT Self-Correction in Radiology Report Simplification [5.059120569845976]
This study explores the suitability of large language models in automatically generating those simplifications.
We propose a new evaluation protocol that employs radiologists and laypeople, where radiologists verify the factual correctness of simplifications.
Our experimental results demonstrate the effectiveness of self-correction prompting in producing high-quality simplifications.
arXiv Detail & Related papers (2024-06-27T03:05:35Z) - Effectiveness of ChatGPT in explaining complex medical reports to patients [3.8249624199073553]
We explore whether ChatGPT (GPT 4) can help explain multidisciplinary team (MDT) reports to colorectal and prostate cancer patients.
These reports are written in dense medical language and assume clinical knowledge, so they are a good test of the ability of ChatGPT to explain complex medical reports to patients.
arXiv Detail & Related papers (2024-06-23T00:04:07Z) - Radiology Report Generation Using Transformers Conditioned with
Non-imaging Data [55.17268696112258]
This paper proposes a novel multi-modal transformer network that integrates chest x-ray (CXR) images and associated patient demographic information.
The proposed network uses a convolutional neural network to extract visual features from CXRs and a transformer-based encoder-decoder network that combines the visual features with semantic text embeddings of patient demographic information.
arXiv Detail & Related papers (2023-11-18T14:52:26Z) - ChatRadio-Valuer: A Chat Large Language Model for Generalizable
Radiology Report Generation Based on Multi-institution and Multi-system Data [115.0747462486285]
ChatRadio-Valuer is a tailored model for automatic radiology report generation that learns generalizable representations.
The clinical dataset utilized in this study encompasses a remarkable total of textbf332,673 observations.
ChatRadio-Valuer consistently outperforms state-of-the-art models, especially ChatGPT (GPT-3.5-Turbo) and GPT-4 et al.
arXiv Detail & Related papers (2023-10-08T17:23:17Z) - Multilingual Natural Language Processing Model for Radiology Reports --
The Summary is all you need! [2.4910932804601855]
The generation of radiology impressions was automated by fine-tuning a model based on a multilingual text-to-text Transformer.
In a blind test, two board-certified radiologists indicated that for at least 70% of the system-generated summaries, the quality matched or exceeded the corresponding human-written summaries.
This study showed that the multilingual model outperformed other models that specialized in summarizing radiology reports in only one language, as well as models that were not specifically designed for summarizing radiology reports.
arXiv Detail & Related papers (2023-09-29T19:20:27Z) - Radiology-GPT: A Large Language Model for Radiology [74.07944784968372]
We introduce Radiology-GPT, a large language model for radiology.
It demonstrates superior performance compared to general language models such as StableLM, Dolly and LLaMA.
It exhibits significant versatility in radiological diagnosis, research, and communication.
arXiv Detail & Related papers (2023-06-14T17:57:24Z) - XrayGPT: Chest Radiographs Summarization using Medical Vision-Language
Models [60.437091462613544]
We introduce XrayGPT, a novel conversational medical vision-language model.
It can analyze and answer open-ended questions about chest radiographs.
We generate 217k interactive and high-quality summaries from free-text radiology reports.
arXiv Detail & Related papers (2023-06-13T17:59:59Z) - To ChatGPT, or not to ChatGPT: That is the question! [78.407861566006]
This study provides a comprehensive and contemporary assessment of the most recent techniques in ChatGPT detection.
We have curated a benchmark dataset consisting of prompts from ChatGPT and humans, including diverse questions from medical, open Q&A, and finance domains.
Our evaluation results demonstrate that none of the existing methods can effectively detect ChatGPT-generated content.
arXiv Detail & Related papers (2023-04-04T03:04:28Z) - Translating Radiology Reports into Plain Language using ChatGPT and
GPT-4 with Prompt Learning: Promising Results, Limitations, and Potential [6.127537348178505]
ChatGPT can successfully translate radiology reports into plain language with an average score of 4.27 in the five-point system.
ChatGPT presents some randomness in its responses with occasionally over-simplified or neglected information.
Results are compared with a newly released large model GPT-4, showing that GPT-4 can significantly improve the quality of reports.
arXiv Detail & Related papers (2023-03-16T02:21:39Z) - Improving Radiology Summarization with Radiograph and Anatomy Prompts [60.30659124918211]
We propose a novel anatomy-enhanced multimodal model to promote impression generation.
In detail, we first construct a set of rules to extract anatomies and put these prompts into each sentence to highlight anatomy characteristics.
We utilize a contrastive learning module to align these two representations at the overall level and use a co-attention to fuse them at the sentence level.
arXiv Detail & Related papers (2022-10-15T14:05:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.