SLPD: Slide-level Prototypical Distillation for WSIs
- URL: http://arxiv.org/abs/2307.10696v1
- Date: Thu, 20 Jul 2023 08:38:15 GMT
- Title: SLPD: Slide-level Prototypical Distillation for WSIs
- Authors: Zhimiao Yu, Tiancheng Lin, Yi Xu
- Abstract summary: We propose Slide-Level Prototypical Distillation (SLPD) to explore intra- and inter-slide semantic structures for context modeling.
SLPD achieves state-of-the-art results on multiple slide-level benchmarks and demonstrates that representation learning of semantic structures of slides can make a suitable proxy task for WSI analysis.
- Score: 11.217079419686472
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Improving the feature representation ability is the foundation of many whole
slide pathological image (WSIs) tasks. Recent works have achieved great success
in pathological-specific self-supervised learning (SSL). However, most of them
only focus on learning patch-level representations, thus there is still a gap
between pretext and slide-level downstream tasks, e.g., subtyping, grading and
staging. Aiming towards slide-level representations, we propose Slide-Level
Prototypical Distillation (SLPD) to explore intra- and inter-slide semantic
structures for context modeling on WSIs. Specifically, we iteratively perform
intra-slide clustering for the regions (4096x4096 patches) within each WSI to
yield the prototypes and encourage the region representations to be closer to
the assigned prototypes. By representing each slide with its prototypes, we
further select similar slides by the set distance of prototypes and assign the
regions by cross-slide prototypes for distillation. SLPD achieves
state-of-the-art results on multiple slide-level benchmarks and demonstrates
that representation learning of semantic structures of slides can make a
suitable proxy task for WSI analysis. Code will be available at
https://github.com/Carboxy/SLPD.
Related papers
- A self-supervised framework for learning whole slide representations [52.774822784847565]
We present Slide Pre-trained Transformers (SPT) for gigapixel-scale self-supervision of whole slide images.
We benchmark SPT visual representations on five diagnostic tasks across three biomedical microscopy datasets.
arXiv Detail & Related papers (2024-02-09T05:05:28Z) - Sub-action Prototype Learning for Point-level Weakly-supervised Temporal
Action Localization [11.777205793663647]
Point-level weakly-supervised temporal action localization (PWTAL) aims to localize actions with only a single timestamp annotation for each action instance.
Existing methods tend to mine dense pseudo labels to alleviate the label sparsity, but overlook the potential sub-action temporal structures, resulting in inferior performance.
We propose a novel sub-action prototype learning framework (SPL-Loc) which comprises Sub-action Prototype Clustering (SPC) and Ordered Prototype Alignment (OPA)
arXiv Detail & Related papers (2023-09-16T17:57:40Z) - BROW: Better featuRes fOr Whole slide image based on self-distillation [19.295596638166536]
Whole slide image (WSI) processing is becoming part of the key components of standard clinical diagnosis for various diseases.
The performance of most WSI-related tasks relies on the efficacy of the backbone which extracts WSI patch feature representations.
We proposed BROW, a foundation model for extracting better feature representations for WSIs, which can be conveniently adapted to downstream tasks without or with slight fine-tuning.
arXiv Detail & Related papers (2023-09-15T09:11:09Z) - ProtoDiv: Prototype-guided Division of Consistent Pseudo-bags for
Whole-slide Image Classification [5.836559246348487]
Pseudo-bag dividing scheme, often crucial for classification performance, is still an open topic worth exploring.
This paper proposes a novel scheme, ProtoDiv, using a bag prototype to guide the division of WSI pseudo-bags.
arXiv Detail & Related papers (2023-04-13T16:27:08Z) - Progressively Dual Prior Guided Few-shot Semantic Segmentation [57.37506990980975]
Few-shot semantic segmentation task aims at performing segmentation in query images with a few annotated support samples.
We propose a progressively dual prior guided few-shot semantic segmentation network.
arXiv Detail & Related papers (2022-11-20T16:19:47Z) - Self-Supervised Pyramid Representation Learning for Multi-Label Visual
Analysis and Beyond [31.36818611460614]
We propose a Self-Supervised Pyramid Learning (SS-PRL) framework.
The proposed SS-PRL is designed to derive pyramid representations at patch levels via learning proper prototypes.
We show that, with our proposed SS-PRL for model pre-training, one can easily adapt and fine-tune the models for a variety of applications.
arXiv Detail & Related papers (2022-08-30T17:57:14Z) - Beyond the Prototype: Divide-and-conquer Proxies for Few-shot
Segmentation [63.910211095033596]
Few-shot segmentation aims to segment unseen-class objects given only a handful of densely labeled samples.
We propose a simple yet versatile framework in the spirit of divide-and-conquer.
Our proposed approach, named divide-and-conquer proxies (DCP), allows for the development of appropriate and reliable information.
arXiv Detail & Related papers (2022-04-21T06:21:14Z) - Dual Prototypical Contrastive Learning for Few-shot Semantic
Segmentation [55.339405417090084]
We propose a dual prototypical contrastive learning approach tailored to the few-shot semantic segmentation (FSS) task.
The main idea is to encourage the prototypes more discriminative by increasing inter-class distance while reducing intra-class distance in prototype feature space.
We demonstrate that the proposed dual contrastive learning approach outperforms state-of-the-art FSS methods on PASCAL-5i and COCO-20i datasets.
arXiv Detail & Related papers (2021-11-09T08:14:50Z) - SCNet: Enhancing Few-Shot Semantic Segmentation by Self-Contrastive
Background Prototypes [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a query image with only a few annotated examples.
Most of advanced solutions exploit a metric learning framework that performs segmentation through matching each pixel to a learned foreground prototype.
This framework suffers from biased classification due to incomplete construction of sample pairs with the foreground prototype only.
arXiv Detail & Related papers (2021-04-19T11:21:47Z) - Part-aware Prototype Network for Few-shot Semantic Segmentation [50.581647306020095]
We propose a novel few-shot semantic segmentation framework based on the prototype representation.
Our key idea is to decompose the holistic class representation into a set of part-aware prototypes.
We develop a novel graph neural network model to generate and enhance the proposed part-aware prototypes.
arXiv Detail & Related papers (2020-07-13T11:03:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.