Self supervised contrastive learning for digital histopathology
- URL: http://arxiv.org/abs/2011.13971v2
- Date: Tue, 7 Sep 2021 20:02:12 GMT
- Title: Self supervised contrastive learning for digital histopathology
- Authors: Ozan Ciga, Tony Xu, Anne L. Martel
- Abstract summary: We use a contrastive self-supervised learning method called SimCLR that achieved state-of-the-art results on natural-scene images.
We find that combining multiple multi-organ datasets with different types of staining and resolution properties improves the quality of the learned features.
Linear classifiers trained on top of the learned features show that networks pretrained on digital histopathology datasets perform better than ImageNet pretrained networks.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised learning has been a long-standing goal of machine learning and
is especially important for medical image analysis, where the learning can
compensate for the scarcity of labeled datasets. A promising subclass of
unsupervised learning is self-supervised learning, which aims to learn salient
features using the raw input as the learning signal. In this paper, we use a
contrastive self-supervised learning method called SimCLR that achieved
state-of-the-art results on natural-scene images and apply this method to
digital histopathology by collecting and pretraining on 57 histopathology
datasets without any labels. We find that combining multiple multi-organ
datasets with different types of staining and resolution properties improves
the quality of the learned features. Furthermore, we find using more images for
pretraining leads to a better performance in multiple downstream tasks. Linear
classifiers trained on top of the learned features show that networks
pretrained on digital histopathology datasets perform better than ImageNet
pretrained networks, boosting task performances by more than 28% in F1 scores
on average. These findings may also be useful when applying newer contrastive
techniques to histopathology data. Pretrained PyTorch models are made publicly
available at https://github.com/ozanciga/self-supervised-histopathology.
Related papers
- Local-to-Global Self-Supervised Representation Learning for Diabetic Retinopathy Grading [0.0]
This research aims to present a novel hybrid learning model using self-supervised learning and knowledge distillation.
In our algorithm, for the first time among all self-supervised learning and knowledge distillation models, the test dataset is 50% larger than the training dataset.
Compared to a similar state-of-the-art model, our results achieved higher accuracy and more effective representation spaces.
arXiv Detail & Related papers (2024-10-01T15:19:16Z) - Are Deep Learning Classification Results Obtained on CT Scans Fair and
Interpretable? [0.0]
Most lung nodule classification papers using deep learning randomly shuffle data and split it into training, validation, and test sets.
In contrast, deep neural networks trained with strict patient-level separation maintain their accuracy rates even when new patient images are tested.
Heat-map visualizations of the activations of the deep neural networks trained with strict patient-level separation indicate a higher degree of focus on the relevant nodules.
arXiv Detail & Related papers (2023-09-22T05:57:25Z) - Performance of GAN-based augmentation for deep learning COVID-19 image
classification [57.1795052451257]
The biggest challenge in the application of deep learning to the medical domain is the availability of training data.
Data augmentation is a typical methodology used in machine learning when confronted with a limited data set.
In this work, a StyleGAN2-ADA model of Generative Adversarial Networks is trained on the limited COVID-19 chest X-ray image set.
arXiv Detail & Related papers (2023-04-18T15:39:58Z) - LifeLonger: A Benchmark for Continual Disease Classification [59.13735398630546]
We introduce LifeLonger, a benchmark for continual disease classification on the MedMNIST collection.
Task and class incremental learning of diseases address the issue of classifying new samples without re-training the models from scratch.
Cross-domain incremental learning addresses the issue of dealing with datasets originating from different institutions while retaining the previously obtained knowledge.
arXiv Detail & Related papers (2022-04-12T12:25:05Z) - Intelligent Masking: Deep Q-Learning for Context Encoding in Medical
Image Analysis [48.02011627390706]
We develop a novel self-supervised approach that occludes targeted regions to improve the pre-training procedure.
We show that training the agent against the prediction model can significantly improve the semantic features extracted for downstream classification tasks.
arXiv Detail & Related papers (2022-03-25T19:05:06Z) - HistoKT: Cross Knowledge Transfer in Computational Pathology [31.14107299224401]
The lack of well-annotated datasets in computational pathology (CPath) obstructs the application of deep learning techniques for classifying medical images.
Most transfer learning research follows a model-centric approach, tuning network parameters to improve transfer results over few datasets.
arXiv Detail & Related papers (2022-01-27T00:34:19Z) - Learning Representations with Contrastive Self-Supervised Learning for
Histopathology Applications [8.69535649683089]
We show how contrastive self-supervised learning can reduce the annotation effort within digital pathology.
Our results pave the way for realizing the full potential of self-supervised learning for histopathology applications.
arXiv Detail & Related papers (2021-12-10T16:08:57Z) - HistoTransfer: Understanding Transfer Learning for Histopathology [9.231495418218813]
We compare the performance of features extracted from networks trained on ImageNet and histopathology data.
We investigate if features learned using more complex networks lead to gain in performance.
arXiv Detail & Related papers (2021-06-13T18:55:23Z) - A Multi-Stage Attentive Transfer Learning Framework for Improving
COVID-19 Diagnosis [49.3704402041314]
We propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis.
Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains.
Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images.
arXiv Detail & Related papers (2021-01-14T01:39:19Z) - Uncovering the structure of clinical EEG signals with self-supervised
learning [64.4754948595556]
Supervised learning paradigms are often limited by the amount of labeled data that is available.
This phenomenon is particularly problematic in clinically-relevant data, such as electroencephalography (EEG)
By extracting information from unlabeled data, it might be possible to reach competitive performance with deep neural networks.
arXiv Detail & Related papers (2020-07-31T14:34:47Z) - 3D medical image segmentation with labeled and unlabeled data using
autoencoders at the example of liver segmentation in CT images [58.720142291102135]
This work investigates the potential of autoencoder-extracted features to improve segmentation with a convolutional neural network.
A convolutional autoencoder was used to extract features from unlabeled data and a multi-scale, fully convolutional CNN was used to perform the target task of 3D liver segmentation in CT images.
arXiv Detail & Related papers (2020-03-17T20:20:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.