XRayGAN: Consistency-preserving Generation of X-ray Images from
Radiology Reports
- URL: http://arxiv.org/abs/2006.10552v1
- Date: Wed, 17 Jun 2020 05:32:14 GMT
- Title: XRayGAN: Consistency-preserving Generation of X-ray Images from
Radiology Reports
- Authors: Xingyi Yang, Nandiraju Gireesh, Eric Xing, Pengtao Xie
- Abstract summary: We develop methods to generate view-consistent, high-fidelity, and high-resolution X-ray images from radiology reports.
This work represents the first one generating consistent and high-resolution X-ray images from radiology reports.
- Score: 19.360283053558604
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To effectively train medical students to become qualified radiologists, a
large number of X-ray images collected from patients with diverse medical
conditions are needed. However, due to data privacy concerns, such images are
typically difficult to obtain. To address this problem, we develop methods to
generate view-consistent, high-fidelity, and high-resolution X-ray images from
radiology reports to facilitate radiology training of medical students. This
task is presented with several challenges. First, from a single report, images
with different views (e.g., frontal, lateral) need to be generated. How to
ensure consistency of these images (i.e., make sure they are about the same
patient)? Second, X-ray images are required to have high resolution. Otherwise,
many details of diseases would be lost. How to generate high-resolutions
images? Third, radiology reports are long and have complicated structure. How
to effectively understand their semantics to generate high-fidelity images that
accurately reflect the contents of the reports? To address these three
challenges, we propose an XRayGAN composed of three modules: (1) a view
consistency network that maximizes the consistency between generated
frontal-view and lateral-view images; (2) a multi-scale conditional GAN that
progressively generates a cascade of images with increasing resolution; (3) a
hierarchical attentional encoder that learns the latent semantics of a
radiology report by capturing its hierarchical linguistic structure and various
levels of clinical importance of words and sentences. Experiments on two
radiology datasets demonstrate the effectiveness of our methods. To our best
knowledge, this work represents the first one generating consistent and
high-resolution X-ray images from radiology reports. The code is available at
https://github.com/UCSD-AI4H/XRayGAN.
Related papers
- X-ray2CTPA: Generating 3D CTPA scans from 2D X-ray conditioning [24.233484690096898]
Chest X-rays or chest radiography (CXR) enables limited imaging compared to computed tomography (CT) scans.
CT scans entail higher costs, greater radiation exposure, and are less accessible than CXRs.
In this work we explore cross-modal translation from a 2D low contrast-resolution X-ray input to a 3D high contrast and spatial-resolutionA scan.
arXiv Detail & Related papers (2024-06-23T13:53:35Z) - Bootstrapping Chest CT Image Understanding by Distilling Knowledge from X-ray Expert Models [17.75505740079875]
We explore the feasibility of leveraging language as a naturally high-quality supervision for chest CT imaging.
We bootstrap the understanding of 3D chest CT images by distilling chest-related diagnostic knowledge from an extensively pre-trained 2D X-ray expert model.
We train our model with over 12,000 pairs of chest CT images and radiology reports.
arXiv Detail & Related papers (2024-04-07T12:17:40Z) - DeViDe: Faceted medical knowledge for improved medical vision-language pre-training [1.6567372257085946]
Vision-language pre-training for chest X-rays has made significant strides, primarily by utilizing paired radiographs and radiology reports.
We propose DeViDe, a transformer-based method that leverages radiographic descriptions from the open web.
DeViDe incorporates three key features for knowledge-augmented vision language alignment: First, a large-language model-based augmentation is employed to homogenise medical knowledge from diverse sources.
In zero-shot settings, DeViDe performs comparably to fully supervised models on external datasets and achieves state-of-the-art results on three large-scale datasets.
arXiv Detail & Related papers (2024-04-04T17:40:06Z) - Automated Chest X-Ray Report Generator Using Multi-Model Deep Learning
Approach [0.0]
The system generates a radiology report by performing the following three steps: image pre-processing, utilizing deep learning models to detect abnormalities, and producing a report.
The proposed system is expected to reduce the workload of radiologists and increase the accuracy of chest X-ray diagnosis.
arXiv Detail & Related papers (2023-09-28T07:57:03Z) - XrayGPT: Chest Radiographs Summarization using Medical Vision-Language
Models [60.437091462613544]
We introduce XrayGPT, a novel conversational medical vision-language model.
It can analyze and answer open-ended questions about chest radiographs.
We generate 217k interactive and high-quality summaries from free-text radiology reports.
arXiv Detail & Related papers (2023-06-13T17:59:59Z) - Act Like a Radiologist: Radiology Report Generation across Anatomical Regions [50.13206214694885]
X-RGen is a radiologist-minded report generation framework across six anatomical regions.
In X-RGen, we seek to mimic the behaviour of human radiologists, breaking them down into four principal phases.
We enhance the recognition capacity of the image encoder by analysing images and reports across various regions.
arXiv Detail & Related papers (2023-05-26T07:12:35Z) - Generative Residual Attention Network for Disease Detection [51.60842580044539]
We present a novel approach for disease generation in X-rays using a conditional generative adversarial learning.
We generate a corresponding radiology image in a target domain while preserving the identity of the patient.
We then use the generated X-ray image in the target domain to augment our training to improve the detection performance.
arXiv Detail & Related papers (2021-10-25T14:15:57Z) - Cross-Modal Contrastive Learning for Abnormality Classification and
Localization in Chest X-rays with Radiomics using a Feedback Loop [63.81818077092879]
We propose an end-to-end semi-supervised cross-modal contrastive learning framework for medical images.
We first apply an image encoder to classify the chest X-rays and to generate the image features.
The radiomic features are then passed through another dedicated encoder to act as the positive sample for the image features generated from the same chest X-ray.
arXiv Detail & Related papers (2021-04-11T09:16:29Z) - XraySyn: Realistic View Synthesis From a Single Radiograph Through CT
Priors [118.27130593216096]
A radiograph visualizes the internal anatomy of a patient through the use of X-ray, which projects 3D information onto a 2D plane.
To the best of our knowledge, this is the first work on radiograph view synthesis.
We show that by gaining an understanding of radiography in 3D space, our method can be applied to radiograph bone extraction and suppression without groundtruth bone labels.
arXiv Detail & Related papers (2020-12-04T05:08:53Z) - Auxiliary Signal-Guided Knowledge Encoder-Decoder for Medical Report
Generation [107.3538598876467]
We propose an Auxiliary Signal-Guided Knowledge-Decoder (ASGK) to mimic radiologists' working patterns.
ASGK integrates internal visual feature fusion and external medical linguistic information to guide medical knowledge transfer and learning.
arXiv Detail & Related papers (2020-06-06T01:00:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.