Show, Describe and Conclude: On Exploiting the Structure Information of
Chest X-Ray Reports
- URL: http://arxiv.org/abs/2004.12274v2
- Date: Thu, 23 Jul 2020 17:44:44 GMT
- Title: Show, Describe and Conclude: On Exploiting the Structure Information of
Chest X-Ray Reports
- Authors: Baoyu Jing, Zeya Wang, Eric Xing
- Abstract summary: Chest X-Ray (CXR) images are commonly used for clinical screening and diagnosis.
The complex structures between and within sections of the reports pose a great challenge to the automatic report generation.
We propose a novel framework that exploits the structure information between and within report sections for generating CXR imaging reports.
- Score: 5.6070625920019825
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Chest X-Ray (CXR) images are commonly used for clinical screening and
diagnosis. Automatically writing reports for these images can considerably
lighten the workload of radiologists for summarizing descriptive findings and
conclusive impressions. The complex structures between and within sections of
the reports pose a great challenge to the automatic report generation.
Specifically, the section Impression is a diagnostic summarization over the
section Findings; and the appearance of normality dominates each section over
that of abnormality. Existing studies rarely explore and consider this
fundamental structure information. In this work, we propose a novel framework
that exploits the structure information between and within report sections for
generating CXR imaging reports. First, we propose a two-stage strategy that
explicitly models the relationship between Findings and Impression. Second, we
design a novel cooperative multi-agent system that implicitly captures the
imbalanced distribution between abnormality and normality. Experiments on two
CXR report datasets show that our method achieves state-of-the-art performance
in terms of various evaluation metrics. Our results expose that the proposed
approach is able to generate high-quality medical reports through integrating
the structure information.
Related papers
- Structural Entities Extraction and Patient Indications Incorporation for Chest X-ray Report Generation [10.46031380503486]
We introduce a novel method, textbfStructural textbfEntities extraction and patient indications textbfIncorporation (SEI) for chest X-ray report generation.
We employ a structural entities extraction (SEE) approach to eliminate presentation-style vocabulary in reports.
We propose a cross-modal fusion network to integrate information from X-ray images, similar historical cases, and patient-specific indications.
arXiv Detail & Related papers (2024-05-23T01:29:47Z) - A Novel Corpus of Annotated Medical Imaging Reports and Information Extraction Results Using BERT-based Language Models [4.023338734079828]
Medical imaging is critical to the diagnosis, surveillance, and treatment of many health conditions.
Radiologists interpret these complex, unstructured images and articulate their assessments through narrative reports that remain largely unstructured.
This unstructured narrative must be converted into a structured semantic representation to facilitate secondary applications such as retrospective analyses or clinical decision support.
arXiv Detail & Related papers (2024-03-27T19:43:45Z) - Radiology Report Generation Using Transformers Conditioned with
Non-imaging Data [55.17268696112258]
This paper proposes a novel multi-modal transformer network that integrates chest x-ray (CXR) images and associated patient demographic information.
The proposed network uses a convolutional neural network to extract visual features from CXRs and a transformer-based encoder-decoder network that combines the visual features with semantic text embeddings of patient demographic information.
arXiv Detail & Related papers (2023-11-18T14:52:26Z) - Beyond Images: An Integrative Multi-modal Approach to Chest X-Ray Report
Generation [47.250147322130545]
Image-to-text radiology report generation aims to automatically produce radiology reports that describe the findings in medical images.
Most existing methods focus solely on the image data, disregarding the other patient information accessible to radiologists.
We present a novel multi-modal deep neural network framework for generating chest X-rays reports by integrating structured patient data, such as vital signs and symptoms, alongside unstructured clinical notes.
arXiv Detail & Related papers (2023-11-18T14:37:53Z) - Controllable Chest X-Ray Report Generation from Longitudinal
Representations [13.151444796296868]
One strategy to speed up reporting is to integrate automated reporting systems.
Previous approaches to automated radiology reporting generally do not provide the prior study as input.
We introduce two novel aspects: (1) longitudinal learning -- we propose a method to align, leverage the current and prior scan information into a joint longitudinal representation which can be provided to the multimodal report generation model; (2) sentence-anatomy dropout -- a training strategy for controllability in which the report generator model is trained to predict only sentences from the original report which correspond to the subset of anatomical regions given as input.
arXiv Detail & Related papers (2023-10-09T17:22:58Z) - Medical Image Captioning via Generative Pretrained Transformers [57.308920993032274]
We combine two language models, the Show-Attend-Tell and the GPT-3, to generate comprehensive and descriptive radiology records.
The proposed model is tested on two medical datasets, the Open-I, MIMIC-CXR, and the general-purpose MS-COCO.
arXiv Detail & Related papers (2022-09-28T10:27:10Z) - Factored Attention and Embedding for Unstructured-view Topic-related
Ultrasound Report Generation [70.7778938191405]
We propose a novel factored attention and embedding model (termed FAE-Gen) for the unstructured-view topic-related ultrasound report generation.
The proposed FAE-Gen mainly consists of two modules, i.e., view-guided factored attention and topic-oriented factored embedding, which capture the homogeneous and heterogeneous morphological characteristic across different views.
arXiv Detail & Related papers (2022-03-12T15:24:03Z) - Variational Topic Inference for Chest X-Ray Report Generation [102.04931207504173]
Report generation for medical imaging promises to reduce workload and assist diagnosis in clinical practice.
Recent work has shown that deep learning models can successfully caption natural images.
We propose variational topic inference for automatic report generation.
arXiv Detail & Related papers (2021-07-15T13:34:38Z) - Learning Visual-Semantic Embeddings for Reporting Abnormal Findings on
Chest X-rays [6.686095511538683]
This work focuses on reporting abnormal findings on radiology images.
We propose a method to identify abnormal findings from the reports in addition to grouping them with unsupervised clustering and minimal rules.
We demonstrate that our method is able to retrieve abnormal findings and outperforms existing generation models on both clinical correctness and text generation metrics.
arXiv Detail & Related papers (2020-10-06T04:18:18Z) - Auxiliary Signal-Guided Knowledge Encoder-Decoder for Medical Report
Generation [107.3538598876467]
We propose an Auxiliary Signal-Guided Knowledge-Decoder (ASGK) to mimic radiologists' working patterns.
ASGK integrates internal visual feature fusion and external medical linguistic information to guide medical knowledge transfer and learning.
arXiv Detail & Related papers (2020-06-06T01:00:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.