Big Self-Supervised Models Advance Medical Image Classification
- URL: http://arxiv.org/abs/2101.05224v1
- Date: Wed, 13 Jan 2021 17:36:31 GMT
- Title: Big Self-Supervised Models Advance Medical Image Classification
- Authors: Shekoofeh Azizi, Basil Mustafa, Fiona Ryan, Zachary Beaver, Jan
Freyberg, Jonathan Deaton, Aaron Loh, Alan Karthikesalingam, Simon Kornblith,
Ting Chen, Vivek Natarajan, Mohammad Norouzi
- Abstract summary: We study the effectiveness of self-supervised learning as a pretraining strategy for medical image classification.
We use a novel Multi-Instance Contrastive Learning (MICLe) method that uses multiple images of the underlying pathology per patient case.
We show that big self-supervised models are robust to distribution shift and can learn efficiently with a small number of labeled medical images.
- Score: 36.23989703428874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-supervised pretraining followed by supervised fine-tuning has seen
success in image recognition, especially when labeled examples are scarce, but
has received limited attention in medical image analysis. This paper studies
the effectiveness of self-supervised learning as a pretraining strategy for
medical image classification. We conduct experiments on two distinct tasks:
dermatology skin condition classification from digital camera images and
multi-label chest X-ray classification, and demonstrate that self-supervised
learning on ImageNet, followed by additional self-supervised learning on
unlabeled domain-specific medical images significantly improves the accuracy of
medical image classifiers. We introduce a novel Multi-Instance Contrastive
Learning (MICLe) method that uses multiple images of the underlying pathology
per patient case, when available, to construct more informative positive pairs
for self-supervised learning. Combining our contributions, we achieve an
improvement of 6.7% in top-1 accuracy and an improvement of 1.1% in mean AUC on
dermatology and chest X-ray classification respectively, outperforming strong
supervised baselines pretrained on ImageNet. In addition, we show that big
self-supervised models are robust to distribution shift and can learn
efficiently with a small number of labeled medical images.
Related papers
- Multi-organ Self-supervised Contrastive Learning for Breast Lesion
Segmentation [0.0]
This paper employs multi-organ datasets for pre-training models tailored to specific organ-related target tasks.
Our target task is breast tumour segmentation in ultrasound images.
Results show that conventional contrastive learning pre-training improves performance compared to supervised baseline approaches.
arXiv Detail & Related papers (2024-02-21T20:29:21Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - GraVIS: Grouping Augmented Views from Independent Sources for
Dermatology Analysis [52.04899592688968]
We propose GraVIS, which is specifically optimized for learning self-supervised features from dermatology images.
GraVIS significantly outperforms its transfer learning and self-supervised learning counterparts in both lesion segmentation and disease classification tasks.
arXiv Detail & Related papers (2023-01-11T11:38:37Z) - RadTex: Learning Efficient Radiograph Representations from Text Reports [7.090896766922791]
We build a data-efficient learning framework that utilizes radiology reports to improve medical image classification performance with limited labeled data.
Our model achieves higher classification performance than ImageNet-supervised pretraining when labeled training data is limited.
arXiv Detail & Related papers (2022-08-05T15:06:26Z) - Performance or Trust? Why Not Both. Deep AUC Maximization with
Self-Supervised Learning for COVID-19 Chest X-ray Classifications [72.52228843498193]
In training deep learning models, a compromise often must be made between performance and trust.
In this work, we integrate a new surrogate loss with self-supervised learning for computer-aided screening of COVID-19 patients.
arXiv Detail & Related papers (2021-12-14T21:16:52Z) - How Transferable Are Self-supervised Features in Medical Image
Classification Tasks? [0.7734726150561086]
Transfer learning has become a standard practice to mitigate the lack of labeled data in medical classification tasks.
Self-supervised pretrained models yield richer embeddings than their supervised counterpart.
Dynamic Visual Meta-Embedding (DVME) is an end-to-end transfer learning approach that fuses pretrained embeddings from multiple models.
arXiv Detail & Related papers (2021-08-23T10:39:31Z) - A Systematic Benchmarking Analysis of Transfer Learning for Medical
Image Analysis [7.339428207644444]
We conduct a systematic study on the transferability of models pre-trained on iNat2021, the most recent large-scale fine-grained dataset.
We present a practical approach to bridge the domain gap between natural and medical images by continually (pre-training) supervised ImageNet models on medical images.
arXiv Detail & Related papers (2021-08-12T19:08:34Z) - Lesion-based Contrastive Learning for Diabetic Retinopathy Grading from
Fundus Images [2.498907460918493]
We propose a self-supervised framework, namely lesion-based contrastive learning for automated diabetic retinopathy grading.
Our proposed framework performs outstandingly on DR grading in terms of both linear evaluation and transfer capacity evaluation.
arXiv Detail & Related papers (2021-07-17T16:30:30Z) - On the Robustness of Pretraining and Self-Supervision for a Deep
Learning-based Analysis of Diabetic Retinopathy [70.71457102672545]
We compare the impact of different training procedures for diabetic retinopathy grading.
We investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions.
Our results indicate that models from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions.
arXiv Detail & Related papers (2021-06-25T08:32:45Z) - Multi-label Thoracic Disease Image Classification with Cross-Attention
Networks [65.37531731899837]
We propose a novel scheme of Cross-Attention Networks (CAN) for automated thoracic disease classification from chest x-ray images.
We also design a new loss function that beyond cross-entropy loss to help cross-attention process and is able to overcome the imbalance between classes and easy-dominated samples within each class.
arXiv Detail & Related papers (2020-07-21T14:37:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.