SLPD: Slide-level Prototypical Distillation for WSIs
- URL: http://arxiv.org/abs/2307.10696v1
- Date: Thu, 20 Jul 2023 08:38:15 GMT
- Title: SLPD: Slide-level Prototypical Distillation for WSIs
- Authors: Zhimiao Yu, Tiancheng Lin, Yi Xu
- Abstract summary: We propose Slide-Level Prototypical Distillation (SLPD) to explore intra- and inter-slide semantic structures for context modeling.
SLPD achieves state-of-the-art results on multiple slide-level benchmarks and demonstrates that representation learning of semantic structures of slides can make a suitable proxy task for WSI analysis.
- Score: 11.217079419686472
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Improving the feature representation ability is the foundation of many whole
slide pathological image (WSIs) tasks. Recent works have achieved great success
in pathological-specific self-supervised learning (SSL). However, most of them
only focus on learning patch-level representations, thus there is still a gap
between pretext and slide-level downstream tasks, e.g., subtyping, grading and
staging. Aiming towards slide-level representations, we propose Slide-Level
Prototypical Distillation (SLPD) to explore intra- and inter-slide semantic
structures for context modeling on WSIs. Specifically, we iteratively perform
intra-slide clustering for the regions (4096x4096 patches) within each WSI to
yield the prototypes and encourage the region representations to be closer to
the assigned prototypes. By representing each slide with its prototypes, we
further select similar slides by the set distance of prototypes and assign the
regions by cross-slide prototypes for distillation. SLPD achieves
state-of-the-art results on multiple slide-level benchmarks and demonstrates
that representation learning of semantic structures of slides can make a
suitable proxy task for WSI analysis. Code will be available at
https://github.com/Carboxy/SLPD.
Related papers
- Cross-Modal Prototype Allocation: Unsupervised Slide Representation Learning via Patch-Text Contrast in Computational Pathology [10.811667603360041]
ProAlign is a cross-modal unsupervised slide representation learning framework.
We leverage a large language model (LLM) to generate descriptive text for the prototype types present in a whole slide image.
We propose a parameter-free attention aggregation strategy that utilizes the similarity between patches and these prototypes to form unsupervised slide embeddings.
arXiv Detail & Related papers (2025-03-26T03:31:07Z) - Slide-Level Prompt Learning with Vision Language Models for Few-Shot Multiple Instance Learning in Histopathology [21.81603581614496]
We address the challenge of few-shot classification in histopathology whole slide images (WSIs)
Our method distinguishes itself by utilizing pathological prior knowledge from language models to identify crucial local tissue types (patches) for WSI classification.
Our approach effectively aligns patch images with tissue types, and we fine-tune our model via prompt learning using only a few labeled WSIs per category.
arXiv Detail & Related papers (2025-03-21T15:40:37Z) - Unsupervised Foundation Model-Agnostic Slide-Level Representation Learning [0.0]
We propose a single modality SSL method in feature space that generates useful slide representations.
Our contrastive pretraining strategy, called COBRA, employs multiple FMs and an architecture based on Mamba-2.
COBRA exceeds performance of state-of-the-art slide encoders on four different public CPTAC cohorts on average by at least +3.8% AUC.
arXiv Detail & Related papers (2024-11-20T13:12:43Z) - A self-supervised framework for learning whole slide representations [52.774822784847565]
We present Slide Pre-trained Transformers (SPT) for gigapixel-scale self-supervision of whole slide images.
We benchmark SPT visual representations on five diagnostic tasks across three biomedical microscopy datasets.
arXiv Detail & Related papers (2024-02-09T05:05:28Z) - BROW: Better featuRes fOr Whole slide image based on self-distillation [19.295596638166536]
Whole slide image (WSI) processing is becoming part of the key components of standard clinical diagnosis for various diseases.
The performance of most WSI-related tasks relies on the efficacy of the backbone which extracts WSI patch feature representations.
We proposed BROW, a foundation model for extracting better feature representations for WSIs, which can be conveniently adapted to downstream tasks without or with slight fine-tuning.
arXiv Detail & Related papers (2023-09-15T09:11:09Z) - ProtoDiv: Prototype-guided Division of Consistent Pseudo-bags for
Whole-slide Image Classification [5.836559246348487]
Pseudo-bag dividing scheme, often crucial for classification performance, is still an open topic worth exploring.
This paper proposes a novel scheme, ProtoDiv, using a bag prototype to guide the division of WSI pseudo-bags.
arXiv Detail & Related papers (2023-04-13T16:27:08Z) - Progressively Dual Prior Guided Few-shot Semantic Segmentation [57.37506990980975]
Few-shot semantic segmentation task aims at performing segmentation in query images with a few annotated support samples.
We propose a progressively dual prior guided few-shot semantic segmentation network.
arXiv Detail & Related papers (2022-11-20T16:19:47Z) - Self-Supervised Pyramid Representation Learning for Multi-Label Visual
Analysis and Beyond [31.36818611460614]
We propose a Self-Supervised Pyramid Learning (SS-PRL) framework.
The proposed SS-PRL is designed to derive pyramid representations at patch levels via learning proper prototypes.
We show that, with our proposed SS-PRL for model pre-training, one can easily adapt and fine-tune the models for a variety of applications.
arXiv Detail & Related papers (2022-08-30T17:57:14Z) - Beyond the Prototype: Divide-and-conquer Proxies for Few-shot
Segmentation [63.910211095033596]
Few-shot segmentation aims to segment unseen-class objects given only a handful of densely labeled samples.
We propose a simple yet versatile framework in the spirit of divide-and-conquer.
Our proposed approach, named divide-and-conquer proxies (DCP), allows for the development of appropriate and reliable information.
arXiv Detail & Related papers (2022-04-21T06:21:14Z) - Dual Prototypical Contrastive Learning for Few-shot Semantic
Segmentation [55.339405417090084]
We propose a dual prototypical contrastive learning approach tailored to the few-shot semantic segmentation (FSS) task.
The main idea is to encourage the prototypes more discriminative by increasing inter-class distance while reducing intra-class distance in prototype feature space.
We demonstrate that the proposed dual contrastive learning approach outperforms state-of-the-art FSS methods on PASCAL-5i and COCO-20i datasets.
arXiv Detail & Related papers (2021-11-09T08:14:50Z) - SCNet: Enhancing Few-Shot Semantic Segmentation by Self-Contrastive
Background Prototypes [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a query image with only a few annotated examples.
Most of advanced solutions exploit a metric learning framework that performs segmentation through matching each pixel to a learned foreground prototype.
This framework suffers from biased classification due to incomplete construction of sample pairs with the foreground prototype only.
arXiv Detail & Related papers (2021-04-19T11:21:47Z) - Part-aware Prototype Network for Few-shot Semantic Segmentation [50.581647306020095]
We propose a novel few-shot semantic segmentation framework based on the prototype representation.
Our key idea is to decompose the holistic class representation into a set of part-aware prototypes.
We develop a novel graph neural network model to generate and enhance the proposed part-aware prototypes.
arXiv Detail & Related papers (2020-07-13T11:03:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.