Auxiliary Signal-Guided Knowledge Encoder-Decoder for Medical Report
Generation
- URL: http://arxiv.org/abs/2006.03744v1
- Date: Sat, 6 Jun 2020 01:00:15 GMT
- Title: Auxiliary Signal-Guided Knowledge Encoder-Decoder for Medical Report
Generation
- Authors: Mingjie Li, Fuyu Wang, Xiaojun Chang and Xiaodan Liang
- Abstract summary: We propose an Auxiliary Signal-Guided Knowledge-Decoder (ASGK) to mimic radiologists' working patterns.
ASGK integrates internal visual feature fusion and external medical linguistic information to guide medical knowledge transfer and learning.
- Score: 107.3538598876467
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Beyond the common difficulties faced in the natural image captioning, medical
report generation specifically requires the model to describe a medical image
with a fine-grained and semantic-coherence paragraph that should satisfy both
medical commonsense and logic. Previous works generally extract the global
image features and attempt to generate a paragraph that is similar to
referenced reports; however, this approach has two limitations. Firstly, the
regions of primary interest to radiologists are usually located in a small area
of the global image, meaning that the remainder parts of the image could be
considered as irrelevant noise in the training procedure. Secondly, there are
many similar sentences used in each medical report to describe the normal
regions of the image, which causes serious data bias. This deviation is likely
to teach models to generate these inessential sentences on a regular basis. To
address these problems, we propose an Auxiliary Signal-Guided Knowledge
Encoder-Decoder (ASGK) to mimic radiologists' working patterns. In more detail,
ASGK integrates internal visual feature fusion and external medical linguistic
information to guide medical knowledge transfer and learning. The core
structure of ASGK consists of a medical graph encoder and a natural language
decoder, inspired by advanced Generative Pre-Training (GPT). Experiments on the
CX-CHR dataset and our COVID-19 CT Report dataset demonstrate that our proposed
ASGK is able to generate a robust and accurate report, and moreover outperforms
state-of-the-art methods on both medical terminology classification and
paragraph generation metrics.
Related papers
- A Novel Corpus of Annotated Medical Imaging Reports and Information Extraction Results Using BERT-based Language Models [4.023338734079828]
Medical imaging is critical to the diagnosis, surveillance, and treatment of many health conditions.
Radiologists interpret these complex, unstructured images and articulate their assessments through narrative reports that remain largely unstructured.
This unstructured narrative must be converted into a structured semantic representation to facilitate secondary applications such as retrospective analyses or clinical decision support.
arXiv Detail & Related papers (2024-03-27T19:43:45Z) - Reading Radiology Imaging Like The Radiologist [3.218449686637963]
We design a factual consistency captioning generator to generate more accurate and factually consistent disease descriptions.
Our framework can find most similar reports for a given disease from the CXR database by retrieving a disease-oriented mask.
arXiv Detail & Related papers (2023-07-12T05:36:47Z) - Replace and Report: NLP Assisted Radiology Report Generation [31.309987297324845]
We propose a template-based approach to generate radiology reports from radiographs.
This is the first attempt to generate chest X-ray radiology reports by first creating small sentences for abnormal findings and then replacing them in the normal report template.
arXiv Detail & Related papers (2023-06-19T10:04:42Z) - Generation of Radiology Findings in Chest X-Ray by Leveraging
Collaborative Knowledge [6.792487817626456]
The cognitive task of interpreting medical images remains the most critical and often time-consuming step in the radiology workflow.
This work focuses on reducing the workload of radiologists who spend most of their time either writing or narrating the Findings.
Unlike past research, which addresses radiology report generation as a single-step image captioning task, we have further taken into consideration the complexity of interpreting CXR images.
arXiv Detail & Related papers (2023-06-18T00:51:28Z) - Act Like a Radiologist: Radiology Report Generation across Anatomical Regions [50.13206214694885]
X-RGen is a radiologist-minded report generation framework across six anatomical regions.
In X-RGen, we seek to mimic the behaviour of human radiologists, breaking them down into four principal phases.
We enhance the recognition capacity of the image encoder by analysing images and reports across various regions.
arXiv Detail & Related papers (2023-05-26T07:12:35Z) - Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report
Generation [92.73584302508907]
We propose a knowledge graph with Dynamic structure and nodes to facilitate medical report generation with Contrastive Learning.
In detail, the fundamental structure of our graph is pre-constructed from general knowledge.
Each image feature is integrated with its very own updated graph before being fed into the decoder module for report generation.
arXiv Detail & Related papers (2023-03-18T03:53:43Z) - Medical Image Captioning via Generative Pretrained Transformers [57.308920993032274]
We combine two language models, the Show-Attend-Tell and the GPT-3, to generate comprehensive and descriptive radiology records.
The proposed model is tested on two medical datasets, the Open-I, MIMIC-CXR, and the general-purpose MS-COCO.
arXiv Detail & Related papers (2022-09-28T10:27:10Z) - Radiomics-Guided Global-Local Transformer for Weakly Supervised
Pathology Localization in Chest X-Rays [65.88435151891369]
Radiomics-Guided Transformer (RGT) fuses textitglobal image information with textitlocal knowledge-guided radiomics information.
RGT consists of an image Transformer branch, a radiomics Transformer branch, and fusion layers that aggregate image and radiomic information.
arXiv Detail & Related papers (2022-07-10T06:32:56Z) - Generative Residual Attention Network for Disease Detection [51.60842580044539]
We present a novel approach for disease generation in X-rays using a conditional generative adversarial learning.
We generate a corresponding radiology image in a target domain while preserving the identity of the patient.
We then use the generated X-ray image in the target domain to augment our training to improve the detection performance.
arXiv Detail & Related papers (2021-10-25T14:15:57Z) - When Radiology Report Generation Meets Knowledge Graph [17.59749125131158]
The accuracy of positive disease keyword mentions is critical in radiology image reporting.
The evaluation of reporting quality should focus more on matching the disease keywords and their associated attributes.
We propose a new evaluation metric for radiology image reporting with the assistance of the same composed graph.
arXiv Detail & Related papers (2020-02-19T16:39:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.