Generalized Radiograph Representation Learning via Cross-supervision
between Images and Free-text Radiology Reports
- URL: http://arxiv.org/abs/2111.03452v1
- Date: Thu, 4 Nov 2021 14:28:22 GMT
- Title: Generalized Radiograph Representation Learning via Cross-supervision
between Images and Free-text Radiology Reports
- Authors: Hong-Yu Zhou, Xiaoyu Chen, Yinghao Zhang, Ruibang Luo, Liansheng Wang,
Yizhou Yu
- Abstract summary: Pre-training lays the foundation for recent successes in radiograph analysis supported by deep learning.
We propose a cross-supervised methodology named REviewing FreE-text Reports for Supervision (REFERS)
REFERS outperforms transfer learning and self-supervised learning counterparts on 4 well-known X-ray datasets under extremely limited supervision.
- Score: 40.42674870179363
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Pre-training lays the foundation for recent successes in radiograph analysis
supported by deep learning. It learns transferable image representations by
conducting large-scale fully-supervised or self-supervised learning on a source
domain. However, supervised pre-training requires a complex and labor intensive
two-stage human-assisted annotation process while self-supervised learning
cannot compete with the supervised paradigm. To tackle these issues, we propose
a cross-supervised methodology named REviewing FreE-text Reports for
Supervision (REFERS), which acquires free supervision signals from original
radiology reports accompanying the radiographs. The proposed approach employs a
vision transformer and is designed to learn joint representations from multiple
views within every patient study. REFERS outperforms its transfer learning and
self-supervised learning counterparts on 4 well-known X-ray datasets under
extremely limited supervision. Moreover, REFERS even surpasses methods based on
a source domain of radiographs with human-assisted structured labels. Thus
REFERS has the potential to replace canonical pre-training methodologies.
Related papers
- Self-supervised pre-training with diffusion model for few-shot landmark detection in x-ray images [0.8793721044482612]
This study introduces a new self-supervised pre-training protocol based on diffusion models for landmark detection in x-ray images.
Our results show that the proposed self-supervised framework can provide accurate landmark detection with a minimal number of available annotated training images.
arXiv Detail & Related papers (2024-07-25T15:32:59Z) - Radiology Report Generation Using Transformers Conditioned with
Non-imaging Data [55.17268696112258]
This paper proposes a novel multi-modal transformer network that integrates chest x-ray (CXR) images and associated patient demographic information.
The proposed network uses a convolutional neural network to extract visual features from CXRs and a transformer-based encoder-decoder network that combines the visual features with semantic text embeddings of patient demographic information.
arXiv Detail & Related papers (2023-11-18T14:52:26Z) - MUSCLE: Multi-task Self-supervised Continual Learning to Pre-train Deep
Models for X-ray Images of Multiple Body Parts [63.30352394004674]
Multi-task Self-super-vised Continual Learning (MUSCLE) is a novel self-supervised pre-training pipeline for medical imaging tasks.
MUSCLE aggregates X-rays collected from multiple body parts for representation learning, and adopts a well-designed continual learning procedure.
We evaluate MUSCLE using 9 real-world X-ray datasets with various tasks, including pneumonia classification, skeletal abnormality classification, lung segmentation, and tuberculosis (TB) detection.
arXiv Detail & Related papers (2023-10-03T12:19:19Z) - Local Contrastive Learning for Medical Image Recognition [0.0]
Local Region Contrastive Learning (LRCLR) is a flexible fine-tuning framework that adds layers for significant image region selection and cross-modality interaction.
Our results on an external validation set of chest x-rays suggest that LRCLR identifies significant local image regions and provides meaningful interpretation against radiology text.
arXiv Detail & Related papers (2023-03-24T17:04:26Z) - Generative Residual Attention Network for Disease Detection [51.60842580044539]
We present a novel approach for disease generation in X-rays using a conditional generative adversarial learning.
We generate a corresponding radiology image in a target domain while preserving the identity of the patient.
We then use the generated X-ray image in the target domain to augment our training to improve the detection performance.
arXiv Detail & Related papers (2021-10-25T14:15:57Z) - A Multi-Stage Attentive Transfer Learning Framework for Improving
COVID-19 Diagnosis [49.3704402041314]
We propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis.
Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains.
Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images.
arXiv Detail & Related papers (2021-01-14T01:39:19Z) - A Multisite, Report-Based, Centralized Infrastructure for Feedback and
Monitoring of Radiology AI/ML Development and Clinical Deployment [0.0]
An interactive radiology reporting approach integrates image viewing, dictation, natural language processing (NLP) and creation of hyperlinks between image findings and the report.
These images and labels can be captured and centralized in a cloud-based system.
The method addresses proposed regulatory requirements for post-marketing surveillance and external data.
arXiv Detail & Related papers (2020-08-31T17:59:04Z) - Joint Modeling of Chest Radiographs and Radiology Reports for Pulmonary
Edema Assessment [39.60171837961607]
We develop a neural network model that is trained on both images and free-text to assess pulmonary edema severity from chest radiographs at inference time.
Our experimental results suggest that the joint image-text representation learning improves the performance of pulmonary edema assessment.
arXiv Detail & Related papers (2020-08-22T17:28:39Z) - Learning Invariant Feature Representation to Improve Generalization
across Chest X-ray Datasets [55.06983249986729]
We show that a deep learning model performing well when tested on the same dataset as training data starts to perform poorly when it is tested on a dataset from a different source.
By employing an adversarial training strategy, we show that a network can be forced to learn a source-invariant representation.
arXiv Detail & Related papers (2020-08-04T07:41:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.