Cyclic Generative Adversarial Networks With Congruent Image-Report
Generation For Explainable Medical Image Analysis
- URL: http://arxiv.org/abs/2211.08424v1
- Date: Wed, 16 Nov 2022 12:41:21 GMT
- Title: Cyclic Generative Adversarial Networks With Congruent Image-Report
Generation For Explainable Medical Image Analysis
- Authors: Dwarikanath Mahapatra
- Abstract summary: We present a novel framework for explainable labeling and interpretation of medical images.
The aim of the work is to generate trustworthy and faithful explanations for the outputs of a model diagnosing chest x-ray images.
- Score: 5.6512908295414
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: We present a novel framework for explainable labeling and interpretation of
medical images. Medical images require specialized professionals for
interpretation, and are explained (typically) via elaborate textual reports.
Different from prior methods that focus on medical report generation from
images or vice-versa, we novelly generate congruent image--report pairs
employing a cyclic-Generative Adversarial Network (cycleGAN); thereby, the
generated report will adequately explain a medical image, while a
report-generated image that effectively characterizes the text visually should
(sufficiently) resemble the original. The aim of the work is to generate
trustworthy and faithful explanations for the outputs of a model diagnosing
chest x-ray images by pointing a human user to similar cases in support of a
diagnostic decision. Apart from enabling transparent medical image labeling and
interpretation, we achieve report and image-based labeling comparable to prior
methods, including state-of-the-art performance in some cases as evidenced by
experiments on the Indiana Chest X-ray dataset
Related papers
- Contrastive Learning with Counterfactual Explanations for Radiology Report Generation [83.30609465252441]
We propose a textbfCountertextbfFactual textbfExplanations-based framework (CoFE) for radiology report generation.
Counterfactual explanations serve as a potent tool for understanding how decisions made by algorithms can be changed by asking what if'' scenarios.
Experiments on two benchmarks demonstrate that leveraging the counterfactual explanations enables CoFE to generate semantically coherent and factually complete reports.
arXiv Detail & Related papers (2024-07-19T17:24:25Z) - Learning Generalized Medical Image Representations through Image-Graph Contrastive Pretraining [11.520404630575749]
We develop an Image-Graph Contrastive Learning framework that pairs chest X-rays with structured report knowledge graphs automatically extracted from radiology notes.
Our approach uniquely encodes the disconnected graph components via a relational graph convolution network and transformer attention.
arXiv Detail & Related papers (2024-05-15T12:27:38Z) - MedCycle: Unpaired Medical Report Generation via Cycle-Consistency [11.190146577567548]
We introduce an innovative approach that eliminates the need for consistent labeling schemas.
This approach is based on cycle-consistent mapping functions that transform image embeddings into report embeddings.
It outperforms state-of-the-art results in unpaired chest X-ray report generation, demonstrating improvements in both language and clinical metrics.
arXiv Detail & Related papers (2024-03-20T09:40:11Z) - VALD-MD: Visual Attribution via Latent Diffusion for Medical Diagnostics [0.0]
Visual attribution in medical imaging seeks to make evident the diagnostically-relevant components of a medical image.
We here present a novel generative visual attribution technique, one that leverages latent diffusion models in combination with domain-specific large language models.
The resulting system also exhibits a range of latent capabilities including zero-shot localized disease induction.
arXiv Detail & Related papers (2024-01-02T19:51:49Z) - Mining Gaze for Contrastive Learning toward Computer-Assisted Diagnosis [61.089776864520594]
We propose eye-tracking as an alternative to text reports for medical images.
By tracking the gaze of radiologists as they read and diagnose medical images, we can understand their visual attention and clinical reasoning.
We introduce the Medical contrastive Gaze Image Pre-training (McGIP) as a plug-and-play module for contrastive learning frameworks.
arXiv Detail & Related papers (2023-12-11T02:27:45Z) - Radiology Report Generation Using Transformers Conditioned with
Non-imaging Data [55.17268696112258]
This paper proposes a novel multi-modal transformer network that integrates chest x-ray (CXR) images and associated patient demographic information.
The proposed network uses a convolutional neural network to extract visual features from CXRs and a transformer-based encoder-decoder network that combines the visual features with semantic text embeddings of patient demographic information.
arXiv Detail & Related papers (2023-11-18T14:52:26Z) - LIMITR: Leveraging Local Information for Medical Image-Text
Representation [17.102338932907294]
This paper focuses on chest X-ray images and their corresponding radiological reports.
It presents a new model that learns a joint X-ray image & report representation.
arXiv Detail & Related papers (2023-03-21T11:20:34Z) - Variational Topic Inference for Chest X-Ray Report Generation [102.04931207504173]
Report generation for medical imaging promises to reduce workload and assist diagnosis in clinical practice.
Recent work has shown that deep learning models can successfully caption natural images.
We propose variational topic inference for automatic report generation.
arXiv Detail & Related papers (2021-07-15T13:34:38Z) - Automated Knee X-ray Report Generation [12.732469371097347]
We propose to take advantage of past radiological exams and formulate a framework capable of learning the correspondence between the images and reports.
We demonstrate how aggregating the image features of individual exams and using them as conditional inputs when training a language generation model results in auto-generated exam reports.
arXiv Detail & Related papers (2021-05-22T11:59:42Z) - Unifying Relational Sentence Generation and Retrieval for Medical Image
Report Composition [142.42920413017163]
Current methods often generate the most common sentences due to dataset bias for individual case.
We propose a novel framework that unifies template retrieval and sentence generation to handle both common and rare abnormality.
arXiv Detail & Related papers (2021-01-09T04:33:27Z) - Auxiliary Signal-Guided Knowledge Encoder-Decoder for Medical Report
Generation [107.3538598876467]
We propose an Auxiliary Signal-Guided Knowledge-Decoder (ASGK) to mimic radiologists' working patterns.
ASGK integrates internal visual feature fusion and external medical linguistic information to guide medical knowledge transfer and learning.
arXiv Detail & Related papers (2020-06-06T01:00:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.