Can we Adopt Self-supervised Pretraining for Chest X-Rays?
- URL: http://arxiv.org/abs/2211.12931v1
- Date: Wed, 23 Nov 2022 13:06:59 GMT
- Title: Can we Adopt Self-supervised Pretraining for Chest X-Rays?
- Authors: Arsh Verma, Makarand Tapaswi
- Abstract summary: Convolutional Neural Networks (CNN) have seen success in identifying pathologies in Chest X-Ray (CXR) images.
In this work, we analyze the utility of pretraining on unlabeled ImageNet or CXR datasets using various algorithms and in multiple settings.
Some findings of our work include: (i) supervised training with labeled ImageNet learns strong representations that are hard to beat; (ii) self-supervised pretraining on ImageNet (1M images) shows performance similar to self-supervised pretraining on a CXR dataset (100K images); and (iii) the CNN trained on supervised
- Score: 10.529356372817386
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Chest radiograph (or Chest X-Ray, CXR) is a popular medical imaging modality
that is used by radiologists across the world to diagnose heart or lung
conditions. Over the last decade, Convolutional Neural Networks (CNN), have
seen success in identifying pathologies in CXR images. Typically, these CNNs
are pretrained on the standard ImageNet classification task, but this assumes
availability of large-scale annotated datasets. In this work, we analyze the
utility of pretraining on unlabeled ImageNet or Chest X-Ray (CXR) datasets
using various algorithms and in multiple settings. Some findings of our work
include: (i) supervised training with labeled ImageNet learns strong
representations that are hard to beat; (ii) self-supervised pretraining on
ImageNet (~1M images) shows performance similar to self-supervised pretraining
on a CXR dataset (~100K images); and (iii) the CNN trained on supervised
ImageNet can be trained further with self-supervised CXR images leading to
improvements, especially when the downstream dataset is on the order of a few
thousand images.
Related papers
- Enhancing Network Initialization for Medical AI Models Using
Large-Scale, Unlabeled Natural Images [1.883452979588382]
Self-supervised learning (SSL) can be applied to chest radiographs to learn robust features.
We tested our approach on over 800,000 chest radiographs from six large global datasets.
arXiv Detail & Related papers (2023-08-15T10:37:13Z) - Vision-Language Modelling For Radiological Imaging and Reports In The
Low Data Regime [70.04389979779195]
This paper explores training medical vision-language models (VLMs) where the visual and language inputs are embedded into a common space.
We explore several candidate methods to improve low-data performance, including adapting generic pre-trained models to novel image and text domains.
Using text-to-image retrieval as a benchmark, we evaluate the performance of these methods with variable sized training datasets of paired chest X-rays and radiological reports.
arXiv Detail & Related papers (2023-03-30T18:20:00Z) - Localization supervision of chest x-ray classifiers using label-specific
eye-tracking annotation [4.8035104863603575]
Eye-tracking (ET) data can be collected in a non-intrusive way during the clinical workflow of a radiologist.
We use ET data recorded from radiologists while dictating CXR reports to train CNNs.
We extract snippets from the ET data by associating them with the dictation of keywords and use them to supervise the localization of abnormalities.
arXiv Detail & Related papers (2022-07-20T09:26:29Z) - Self-Supervised Learning as a Means To Reduce the Need for Labeled Data
in Medical Image Analysis [64.4093648042484]
We use a dataset of chest X-ray images with bounding box labels for 13 different classes of anomalies.
We show that it is possible to achieve similar performance to a fully supervised model in terms of mean average precision and accuracy with only 60% of the labeled data.
arXiv Detail & Related papers (2022-06-01T09:20:30Z) - Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification [74.84221280249876]
An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
arXiv Detail & Related papers (2022-05-08T15:29:54Z) - An Empirical Study of Remote Sensing Pretraining [117.90699699469639]
We conduct an empirical study of remote sensing pretraining (RSP) on aerial images.
RSP can help deliver distinctive performances in scene recognition tasks.
RSP mitigates the data discrepancies of traditional ImageNet pretraining on RS images, but it may still suffer from task discrepancies.
arXiv Detail & Related papers (2022-04-06T13:38:11Z) - Follow My Eye: Using Gaze to Supervise Computer-Aided Diagnosis [54.60796004113496]
We demonstrate that the eye movement of radiologists reading medical images can be a new form of supervision to train the DNN-based computer-aided diagnosis (CAD) system.
We record the tracks of the radiologists' gaze when they are reading images.
The gaze information is processed and then used to supervise the DNN's attention via an Attention Consistency module.
arXiv Detail & Related papers (2022-04-06T08:31:05Z) - Corrupted Image Modeling for Self-Supervised Visual Pre-Training [103.99311611776697]
We introduce Corrupted Image Modeling (CIM) for self-supervised visual pre-training.
CIM uses an auxiliary generator with a small trainable BEiT to corrupt the input image instead of using artificial mask tokens.
After pre-training, the enhancer can be used as a high-capacity visual encoder for downstream tasks.
arXiv Detail & Related papers (2022-02-07T17:59:04Z) - Multi-Label Generalized Zero Shot Learning for the Classification of
Disease in Chest Radiographs [0.7734726150561088]
We propose a zero shot learning network that can simultaneously predict multiple seen and unseen diseases in chest X-ray images.
The network is end-to-end trainable and requires no independent pre-training for the offline feature extractor.
Our network outperforms two strong baselines in terms of recall, precision, f1 score, and area under the receiver operating characteristic curve.
arXiv Detail & Related papers (2021-07-14T09:04:20Z) - Image Embedding and Model Ensembling for Automated Chest X-Ray
Interpretation [0.0]
We present and study several machine learning approaches to develop automated Chest X-ray diagnostic models.
In particular, we trained several Convolutional Neural Networks (CNN) on the CheXpert dataset.
We used the trained CNNs to compute embeddings of the CXR images, in order to train two sets of tree-based classifiers from them.
arXiv Detail & Related papers (2021-05-05T14:48:59Z) - CVR-Net: A deep convolutional neural network for coronavirus recognition
from chest radiography images [1.869097450593631]
We propose a robust CNN-based network, called CVR-Net, for the automatic recognition of the coronavirus from CT or X-ray images.
We train and test the proposed CVR-Net on three different datasets, where the images have collected from different open-source repositories.
Our model achieves an overall F1-score & accuracy of 0.997 & 0.998; 0.963 & 0.964; 0.816 & 0.820; 0.961 & 0.961; and 0.780 & 0.780, respectively, for task-1 to task-5.
arXiv Detail & Related papers (2020-07-21T18:21:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.