Giga-SSL: Self-Supervised Learning for Gigapixel Images
- URL: http://arxiv.org/abs/2212.03273v1
- Date: Tue, 6 Dec 2022 19:09:19 GMT
- Title: Giga-SSL: Self-Supervised Learning for Gigapixel Images
- Authors: Tristan Lazard, Marvin Lerousseau, Etienne Decenci\`ere, Thomas Walter
- Abstract summary: Whole slide images (WSI) are microscopy images of stained tissue slides routinely prepared for diagnosis and treatment selection in medical practice.
The current state-of-the-art (SoTA) approach to classify WSI subdivides them into tiles, encodes them by pre-trained networks and applies Multiple Instance Learning (MIL) to train for specific downstream tasks.
Here, we propose a strategy of slide level SSL to leverage the large number of WSI without annotations to infer powerful slide representations.
- Score: 0.8029049649310211
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Whole slide images (WSI) are microscopy images of stained tissue slides
routinely prepared for diagnosis and treatment selection in medical practice.
WSI are very large (gigapixel size) and complex (made of up to millions of
cells). The current state-of-the-art (SoTA) approach to classify WSI subdivides
them into tiles, encodes them by pre-trained networks and applies Multiple
Instance Learning (MIL) to train for specific downstream tasks. However,
annotated datasets are often small, typically a few hundred to a few thousand
WSI, which may cause overfitting and underperforming models. Conversely, the
number of unannotated WSI is ever increasing, with datasets of tens of
thousands (soon to be millions) of images available. While it has been
previously proposed to use these unannotated data to identify suitable tile
representations by self-supervised learning (SSL), downstream classification
tasks still require full supervision because parts of the MIL architecture is
not trained during tile level SSL pre-training. Here, we propose a strategy of
slide level SSL to leverage the large number of WSI without annotations to
infer powerful slide representations. Applying our method to The Cancer-Genome
Atlas, one of the most widely used data resources in cancer research (16 TB
image data), we are able to downsize the dataset to 23 MB without any loss in
predictive power: we show that a linear classifier trained on top of these
embeddings maintains or improves previous SoTA performances on various
benchmark WSI classification tasks. Finally, we observe that training a
classifier on these representations with tiny datasets (e.g. 50 slides)
improved performances over SoTA by an average of +6.3 AUC points over all
downstream tasks.
Related papers
- A self-supervised framework for learning whole slide representations [52.774822784847565]
We present Slide Pre-trained Transformers (SPT) for gigapixel-scale self-supervision of whole slide images.
We benchmark SPT visual representations on five diagnostic tasks across three biomedical microscopy datasets.
arXiv Detail & Related papers (2024-02-09T05:05:28Z) - LESS: Label-efficient Multi-scale Learning for Cytological Whole Slide
Image Screening [19.803614403803962]
We propose a weakly-supervised Label-Efficient WSI Screening method, dubbed LESS, for cytological WSI analysis with only slide-level labels.
We provide appropriate supervision by using slide-level labels to improve the learning of patch-level features.
It outperforms state-of-the-art MIL methods on pathology WSIs and realizes automatic cytological WSI cancer screening.
arXiv Detail & Related papers (2023-06-06T05:09:20Z) - CSP: Self-Supervised Contrastive Spatial Pre-Training for
Geospatial-Visual Representations [90.50864830038202]
We present Contrastive Spatial Pre-Training (CSP), a self-supervised learning framework for geo-tagged images.
We use a dual-encoder to separately encode the images and their corresponding geo-locations, and use contrastive objectives to learn effective location representations from images.
CSP significantly boosts the model performance with 10-34% relative improvement with various labeled training data sampling ratios.
arXiv Detail & Related papers (2023-05-01T23:11:18Z) - A Dual-branch Self-supervised Representation Learning Framework for
Tumour Segmentation in Whole Slide Images [12.961686610789416]
Self-supervised learning (SSL) has emerged as an alternative solution to reduce the annotation overheads in whole slide images.
These SSL approaches are not designed for handling multi-resolution WSIs, which limits their performance in learning discriminative image features.
We propose a Dual-branch SSL Framework for WSI tumour segmentation (DSF-WSI) that can effectively learn image features from multi-resolution WSIs.
arXiv Detail & Related papers (2023-03-20T10:57:28Z) - Task-specific Fine-tuning via Variational Information Bottleneck for
Weakly-supervised Pathology Whole Slide Image Classification [10.243293283318415]
Multiple Instance Learning (MIL) has shown promising results in digital Pathology Whole Slide Image (WSI) classification.
We propose an efficient WSI fine-tuning framework motivated by the Information Bottleneck theory.
Our framework is evaluated on five pathology WSI datasets on various WSI heads.
arXiv Detail & Related papers (2023-03-15T08:41:57Z) - Self Supervised Learning for Few Shot Hyperspectral Image Classification [57.2348804884321]
We propose to leverage Self Supervised Learning (SSL) for HSI classification.
We show that by pre-training an encoder on unlabeled pixels using Barlow-Twins, a state-of-the-art SSL algorithm, we can obtain accurate models with a handful of labels.
arXiv Detail & Related papers (2022-06-24T07:21:53Z) - DATA: Domain-Aware and Task-Aware Pre-training [94.62676913928831]
We present DATA, a simple yet effective NAS approach specialized for self-supervised learning (SSL)
Our method achieves promising results across a wide range of computation costs on downstream tasks, including image classification, object detection and semantic segmentation.
arXiv Detail & Related papers (2022-03-17T02:38:49Z) - CvS: Classification via Segmentation For Small Datasets [52.821178654631254]
This paper presents CvS, a cost-effective classifier for small datasets that derives the classification labels from predicting the segmentation maps.
We evaluate the effectiveness of our framework on diverse problems showing that CvS is able to achieve much higher classification results compared to previous methods when given only a handful of examples.
arXiv Detail & Related papers (2021-10-29T18:41:15Z) - Hierarchical Self-Supervised Learning for Medical Image Segmentation
Based on Multi-Domain Data Aggregation [23.616336382437275]
We propose Hierarchical Self-Supervised Learning (HSSL) for medical image segmentation.
We first aggregate a dataset from several medical challenges, then pre-train the network in a self-supervised manner, and finally fine-tune on labeled data.
Compared to learning from scratch, our new method yields better performance on various tasks.
arXiv Detail & Related papers (2021-07-10T18:17:57Z) - TAFSSL: Task-Adaptive Feature Sub-Space Learning for few-shot
classification [50.358839666165764]
We show that the Task-Adaptive Feature Sub-Space Learning (TAFSSL) can significantly boost the performance in Few-Shot Learning scenarios.
Specifically, we show that on the challenging miniImageNet and tieredImageNet benchmarks, TAFSSL can improve the current state-of-the-art in both transductive and semi-supervised FSL settings by more than $5%$.
arXiv Detail & Related papers (2020-03-14T16:59:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.