Self-supervised Learning from 100 Million Medical Images
- URL: http://arxiv.org/abs/2201.01283v1
- Date: Tue, 4 Jan 2022 18:27:04 GMT
- Title: Self-supervised Learning from 100 Million Medical Images
- Authors: Florin C. Ghesu, Bogdan Georgescu, Awais Mansoor, Youngjin Yoo,
Dominik Neumann, Pragneshkumar Patel, R.S. Vishwanath, James M. Balter, Yue
Cao, Sasa Grbic, Dorin Comaniciu
- Abstract summary: We propose a method for self-supervised learning of rich image features based on contrastive learning and online feature clustering.
We leverage large training datasets of over 100,000,000 medical images of various modalities, including radiography, computed tomography (CT), magnetic resonance (MR) imaging and ultrasonography.
We highlight a number of advantages of this strategy on challenging image assessment problems in radiography, CT and MR.
- Score: 13.958840691105992
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Building accurate and robust artificial intelligence systems for medical
image assessment requires not only the research and design of advanced deep
learning models but also the creation of large and curated sets of annotated
training examples. Constructing such datasets, however, is often very costly --
due to the complex nature of annotation tasks and the high level of expertise
required for the interpretation of medical images (e.g., expert radiologists).
To counter this limitation, we propose a method for self-supervised learning of
rich image features based on contrastive learning and online feature
clustering. For this purpose we leverage large training datasets of over
100,000,000 medical images of various modalities, including radiography,
computed tomography (CT), magnetic resonance (MR) imaging and ultrasonography.
We propose to use these features to guide model training in supervised and
hybrid self-supervised/supervised regime on various downstream tasks. We
highlight a number of advantages of this strategy on challenging image
assessment problems in radiography, CT and MR: 1) Significant increase in
accuracy compared to the state-of-the-art (e.g., AUC boost of 3-7% for
detection of abnormalities from chest radiography scans and hemorrhage
detection on brain CT); 2) Acceleration of model convergence during training by
up to 85% compared to using no pretraining (e.g., 83% when training a model for
detection of brain metastases in MR scans); 3) Increase in robustness to
various image augmentations, such as intensity variations, rotations or scaling
reflective of data variation seen in the field.
Related papers
- Towards a vision foundation model for comprehensive assessment of Cardiac MRI [11.838157772803282]
We introduce a vision foundation model trained for cardiac magnetic resonance imaging (CMR) assessment.
We finetune the model in supervised way for 9 clinical tasks typical to a CMR workflow.
We demonstrate improved accuracy and robustness across all tasks, over a range of available labeled dataset sizes.
arXiv Detail & Related papers (2024-10-02T15:32:01Z) - 3D-CT-GPT: Generating 3D Radiology Reports through Integration of Large Vision-Language Models [51.855377054763345]
This paper introduces 3D-CT-GPT, a Visual Question Answering (VQA)-based medical visual language model for generating radiology reports from 3D CT scans.
Experiments on both public and private datasets demonstrate that 3D-CT-GPT significantly outperforms existing methods in terms of report accuracy and quality.
arXiv Detail & Related papers (2024-09-28T12:31:07Z) - Autoregressive Sequence Modeling for 3D Medical Image Representation [48.706230961589924]
We introduce a pioneering method for learning 3D medical image representations through an autoregressive sequence pre-training framework.
Our approach various 3D medical images based on spatial, contrast, and semantic correlations, treating them as interconnected visual tokens within a token sequence.
arXiv Detail & Related papers (2024-09-13T10:19:10Z) - Detecting Brain Tumors through Multimodal Neural Networks [0.0]
This research work aims to assess the performance of a multimodal model for the classification of Magnetic Resonance Imaging (MRI) scans processed as grayscale images.
The results are promising, and in line with similar works, as the model reaches an accuracy of around 98%.
arXiv Detail & Related papers (2024-01-10T13:06:52Z) - On Sensitivity and Robustness of Normalization Schemes to Input
Distribution Shifts in Automatic MR Image Diagnosis [58.634791552376235]
Deep Learning (DL) models have achieved state-of-the-art performance in diagnosing multiple diseases using reconstructed images as input.
DL models are sensitive to varying artifacts as it leads to changes in the input data distribution between the training and testing phases.
We propose to use other normalization techniques, such as Group Normalization and Layer Normalization, to inject robustness into model performance against varying image artifacts.
arXiv Detail & Related papers (2023-06-23T03:09:03Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Iterative Data Refinement for Self-Supervised MR Image Reconstruction [18.02961646651716]
We propose a data refinement framework for self-supervised MR image reconstruction.
We first analyze the reason of the performance gap between self-supervised and supervised methods.
Then, we design an effective self-supervised training data refinement method to reduce this data bias.
arXiv Detail & Related papers (2022-11-24T06:57:16Z) - Deep Learning models for benign and malign Ocular Tumor Growth
Estimation [3.1558405181807574]
Clinicians often face issues in selecting suitable image processing algorithm for medical imaging data.
A strategy for the selection of a proper model is presented here.
arXiv Detail & Related papers (2021-07-09T05:40:25Z) - Imbalance-Aware Self-Supervised Learning for 3D Radiomic Representations [5.750111443935516]
We show how to learn image representations in a self-supervised fashion using a 3D Siamese network.
We show significant improvement in brain tumor classification and lung cancer staging tasks covering MRI and CT imaging modalities.
arXiv Detail & Related papers (2021-03-06T18:17:03Z) - A Multi-Stage Attentive Transfer Learning Framework for Improving
COVID-19 Diagnosis [49.3704402041314]
We propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis.
Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains.
Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images.
arXiv Detail & Related papers (2021-01-14T01:39:19Z) - Generative Adversarial U-Net for Domain-free Medical Image Augmentation [49.72048151146307]
The shortage of annotated medical images is one of the biggest challenges in the field of medical image computing.
In this paper, we develop a novel generative method named generative adversarial U-Net.
Our newly designed model is domain-free and generalizable to various medical images.
arXiv Detail & Related papers (2021-01-12T23:02:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.