Unifying Relational Sentence Generation and Retrieval for Medical Image
Report Composition
- URL: http://arxiv.org/abs/2101.03287v1
- Date: Sat, 9 Jan 2021 04:33:27 GMT
- Title: Unifying Relational Sentence Generation and Retrieval for Medical Image
Report Composition
- Authors: Fuyu Wang and Xiaodan Liang and Lin Xu and Liang Lin
- Abstract summary: Current methods often generate the most common sentences due to dataset bias for individual case.
We propose a novel framework that unifies template retrieval and sentence generation to handle both common and rare abnormality.
- Score: 142.42920413017163
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Beyond generating long and topic-coherent paragraphs in traditional
captioning tasks, the medical image report composition task poses more
task-oriented challenges by requiring both the highly-accurate medical term
diagnosis and multiple heterogeneous forms of information including impression
and findings. Current methods often generate the most common sentences due to
dataset bias for individual case, regardless of whether the sentences properly
capture key entities and relationships. Such limitations severely hinder their
applicability and generalization capability in medical report composition where
the most critical sentences lie in the descriptions of abnormal diseases that
are relatively rare. Moreover, some medical terms appearing in one report are
often entangled with each other and co-occurred, e.g. symptoms associated with
a specific disease. To enforce the semantic consistency of medical terms to be
incorporated into the final reports and encourage the sentence generation for
rare abnormal descriptions, we propose a novel framework that unifies template
retrieval and sentence generation to handle both common and rare abnormality
while ensuring the semantic-coherency among the detected medical terms.
Specifically, our approach exploits hybrid-knowledge co-reasoning: i) explicit
relationships among all abnormal medical terms to induce the visual attention
learning and topic representation encoding for better topic-oriented symptoms
descriptions; ii) adaptive generation mode that changes between the template
retrieval and sentence generation according to a contextual topic encoder.
Experimental results on two medical report benchmarks demonstrate the
superiority of the proposed framework in terms of both human and metrics
evaluation.
Related papers
- Contrastive Learning with Counterfactual Explanations for Radiology Report Generation [83.30609465252441]
We propose a textbfCountertextbfFactual textbfExplanations-based framework (CoFE) for radiology report generation.
Counterfactual explanations serve as a potent tool for understanding how decisions made by algorithms can be changed by asking what if'' scenarios.
Experiments on two benchmarks demonstrate that leveraging the counterfactual explanations enables CoFE to generate semantically coherent and factually complete reports.
arXiv Detail & Related papers (2024-07-19T17:24:25Z) - Topicwise Separable Sentence Retrieval for Medical Report Generation [41.812337937025084]
We introduce a Topicwise Separable Sentence Retrieval (Teaser) for medical report generation.
To ensure comprehensive learning of both common and rare topics, we categorize queries into common and rare types, and then propose Topic Contrastive Loss.
Experiments on the MIMIC-CXR and IU X-ray datasets demonstrate that Teaser surpasses state-of-the-art models.
arXiv Detail & Related papers (2024-05-07T10:21:23Z) - C^2M-DoT: Cross-modal consistent multi-view medical report generation
with domain transfer network [67.97926983664676]
We propose a cross-modal consistent multi-view medical report generation with a domain transfer network (C2M-DoT)
C2M-DoT substantially outperforms state-of-the-art baselines in all metrics.
arXiv Detail & Related papers (2023-10-09T02:31:36Z) - "Nothing Abnormal": Disambiguating Medical Reports via Contrastive
Knowledge Infusion [6.9551174393701345]
We propose a rewriting algorithm based on contrastive pretraining and perturbation-based rewriting.
We create two datasets, OpenI-Annotated based on chest reports and VA-Annotated based on general medical reports.
Our proposed algorithm effectively rewrites input sentences in a less ambiguous way with high content fidelity.
arXiv Detail & Related papers (2023-05-15T02:01:20Z) - A Self-Guided Framework for Radiology Report Generation [10.573538773141715]
A self-guided framework (SGF) is developed to generate medical reports with annotated disease labels.
SGF uses unsupervised and supervised deep learning methods to mimic the process of human learning and writing.
Our results highlight the capacity of the proposed framework to distinguish fined-grained visual details between words.
arXiv Detail & Related papers (2022-06-19T11:09:27Z) - Weakly Supervised Contrastive Learning for Chest X-Ray Report Generation [3.3978173451092437]
Radiology report generation aims at generating descriptive text from radiology images automatically.
A typical setting consists of training encoder-decoder models on image-report pairs with a cross entropy loss.
We propose a novel weakly supervised contrastive loss for medical report generation.
arXiv Detail & Related papers (2021-09-25T00:06:23Z) - Variational Topic Inference for Chest X-Ray Report Generation [102.04931207504173]
Report generation for medical imaging promises to reduce workload and assist diagnosis in clinical practice.
Recent work has shown that deep learning models can successfully caption natural images.
We propose variational topic inference for automatic report generation.
arXiv Detail & Related papers (2021-07-15T13:34:38Z) - Auxiliary Signal-Guided Knowledge Encoder-Decoder for Medical Report
Generation [107.3538598876467]
We propose an Auxiliary Signal-Guided Knowledge-Decoder (ASGK) to mimic radiologists' working patterns.
ASGK integrates internal visual feature fusion and external medical linguistic information to guide medical knowledge transfer and learning.
arXiv Detail & Related papers (2020-06-06T01:00:15Z) - Semi-supervised Medical Image Classification with Relation-driven
Self-ensembling Model [71.80319052891817]
We present a relation-driven semi-supervised framework for medical image classification.
It exploits the unlabeled data by encouraging the prediction consistency of given input under perturbations.
Our method outperforms many state-of-the-art semi-supervised learning methods on both single-label and multi-label image classification scenarios.
arXiv Detail & Related papers (2020-05-15T06:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.