Cluster-to-Conquer: A Framework for End-to-End Multi-Instance Learning
for Whole Slide Image Classification
- URL: http://arxiv.org/abs/2103.10626v1
- Date: Fri, 19 Mar 2021 04:24:01 GMT
- Title: Cluster-to-Conquer: A Framework for End-to-End Multi-Instance Learning
for Whole Slide Image Classification
- Authors: Yash Sharma, Aman Shrivastava, Lubaina Ehsan, Christopher A. Moskaluk,
Sana Syed, Donald E. Brown
- Abstract summary: We propose an end-to-end framework that clusters the patches from a Whole Slide Images (WSI) into $k$-groups, samples $k'$ patches from each group for training, and uses an adaptive attention mechanism for slide level prediction.
The framework is optimized end-to-end on slide-level cross-entropy, patch-level cross-entropy, and KL-divergence loss.
- Score: 7.876654642325896
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, the availability of digitized Whole Slide Images (WSIs) has
enabled the use of deep learning-based computer vision techniques for automated
disease diagnosis. However, WSIs present unique computational and algorithmic
challenges. WSIs are gigapixel-sized ($\sim$100K pixels), making them
infeasible to be used directly for training deep neural networks. Also, often
only slide-level labels are available for training as detailed annotations are
tedious and can be time-consuming for experts. Approaches using
multiple-instance learning (MIL) frameworks have been shown to overcome these
challenges. Current state-of-the-art approaches divide the learning framework
into two decoupled parts: a convolutional neural network (CNN) for encoding the
patches followed by an independent aggregation approach for slide-level
prediction. In this approach, the aggregation step has no bearing on the
representations learned by the CNN encoder. We have proposed an end-to-end
framework that clusters the patches from a WSI into ${k}$-groups, samples
${k}'$ patches from each group for training, and uses an adaptive attention
mechanism for slide level prediction; Cluster-to-Conquer (C2C). We have
demonstrated that dividing a WSI into clusters can improve the model training
by exposing it to diverse discriminative features extracted from the patches.
We regularized the clustering mechanism by introducing a KL-divergence loss
between the attention weights of patches in a cluster and the uniform
distribution. The framework is optimized end-to-end on slide-level
cross-entropy, patch-level cross-entropy, and KL-divergence loss
(Implementation: https://github.com/YashSharma/C2C).
Related papers
- MUSTANG: Multi-Stain Self-Attention Graph Multiple Instance Learning
Pipeline for Histopathology Whole Slide Images [1.127806343149511]
Whole Slide Images (WSIs) present a challenging computer vision task due to their gigapixel size and presence of artefacts.
Real-world clinical datasets tend to come as sets of heterogeneous WSIs with labels present at the patient-level, with poor to no annotations.
Here we propose an end-to-end multi-stain self-attention graph (MUSTANG) multiple instance learning pipeline.
arXiv Detail & Related papers (2023-09-19T14:30:14Z) - Neural Collapse Inspired Feature-Classifier Alignment for Few-Shot Class
Incremental Learning [120.53458753007851]
Few-shot class-incremental learning (FSCIL) has been a challenging problem as only a few training samples are accessible for each novel class in the new sessions.
We deal with this misalignment dilemma in FSCIL inspired by the recently discovered phenomenon named neural collapse.
We propose a neural collapse inspired framework for FSCIL. Experiments on the miniImageNet, CUB-200, and CIFAR-100 datasets demonstrate that our proposed framework outperforms the state-of-the-art performances.
arXiv Detail & Related papers (2023-02-06T18:39:40Z) - Scale Attention for Learning Deep Face Representation: A Study Against
Visual Scale Variation [69.45176408639483]
We reform the conv layer by resorting to the scale-space theory.
We build a novel style named SCale AttentioN Conv Neural Network (textbfSCAN-CNN)
As a single-shot scheme, the inference is more efficient than multi-shot fusion.
arXiv Detail & Related papers (2022-09-19T06:35:04Z) - Gigapixel Whole-Slide Images Classification using Locally Supervised
Learning [31.213316201151954]
Histo whole slide images (WSIs) play a very important role in clinical studies and serve as the gold standard for many cancer diagnoses.
Conventional methods rely on a multiple instance learning (MIL) strategy to process a WSI at patch level.
We propose a locally supervised learning framework which processes the entire slide by exploring the entire local and global information.
arXiv Detail & Related papers (2022-07-17T19:31:54Z) - Deep Image Clustering with Contrastive Learning and Multi-scale Graph
Convolutional Networks [58.868899595936476]
This paper presents a new deep clustering approach termed image clustering with contrastive learning and multi-scale graph convolutional networks (IcicleGCN)
Experiments on multiple image datasets demonstrate the superior clustering performance of IcicleGCN over the state-of-the-art.
arXiv Detail & Related papers (2022-07-14T19:16:56Z) - Decoupled Multi-task Learning with Cyclical Self-Regulation for Face
Parsing [71.19528222206088]
We propose a novel Decoupled Multi-task Learning with Cyclical Self-Regulation for face parsing.
Specifically, DML-CSR designs a multi-task model which comprises face parsing, binary edge, and category edge detection.
Our method achieves the new state-of-the-art performance on the Helen, CelebA-HQ, and LapaMask datasets.
arXiv Detail & Related papers (2022-03-28T02:12:30Z) - Accounting for Dependencies in Deep Learning Based Multiple Instance
Learning for Whole Slide Imaging [8.712556146101953]
Multiple instance learning (MIL) is a key algorithm for classification of whole slide images (WSI)
Histology WSIs can have billions of pixels, which create enormous computational and annotation challenges.
We propose an instance-wise loss function based on instance pseudo-labels.
arXiv Detail & Related papers (2021-11-01T06:50:33Z) - Learning Hierarchical Graph Neural Networks for Image Clustering [81.5841862489509]
We propose a hierarchical graph neural network (GNN) model that learns how to cluster a set of images into an unknown number of identities.
Our hierarchical GNN uses a novel approach to merge connected components predicted at each level of the hierarchy to form a new graph at the next level.
arXiv Detail & Related papers (2021-07-03T01:28:42Z) - Pay Attention with Focus: A Novel Learning Scheme for Classification of
Whole Slide Images [8.416553728391309]
We propose a novel two-stage approach to analyze whole slide images (WSIs)
First, we extract a set of representative patches (called mosaic) from a WSI.
Each patch of a mosaic is encoded to a feature vector using a deep network.
In the second stage, a set of encoded patch-level features from a WSI is used to compute the primary diagnosis probability.
arXiv Detail & Related papers (2021-06-11T21:59:02Z) - Embedding Propagation: Smoother Manifold for Few-Shot Classification [131.81692677836202]
We propose to use embedding propagation as an unsupervised non-parametric regularizer for manifold smoothing in few-shot classification.
We empirically show that embedding propagation yields a smoother embedding manifold.
We show that embedding propagation consistently improves the accuracy of the models in multiple semi-supervised learning scenarios by up to 16% points.
arXiv Detail & Related papers (2020-03-09T13:51:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.