ReMix: A General and Efficient Framework for Multiple Instance Learning
based Whole Slide Image Classification
- URL: http://arxiv.org/abs/2207.01805v1
- Date: Tue, 5 Jul 2022 04:21:35 GMT
- Title: ReMix: A General and Efficient Framework for Multiple Instance Learning
based Whole Slide Image Classification
- Authors: Jiawei Yang, Hanbo Chen, Yu Zhao, Fan Yang, Yao Zhang, Lei He, Jianhua
Yao
- Abstract summary: Whole slide image (WSI) classification often relies on weakly supervised multiple instance learning (MIL) methods to handle gigapixel resolution images and slide-level labels.
We propose ReMix, a general and efficient framework for MIL based WSI classification.
- Score: 14.78430890440035
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Whole slide image (WSI) classification often relies on deep weakly supervised
multiple instance learning (MIL) methods to handle gigapixel resolution images
and slide-level labels. Yet the decent performance of deep learning comes from
harnessing massive datasets and diverse samples, urging the need for efficient
training pipelines for scaling to large datasets and data augmentation
techniques for diversifying samples. However, current MIL-based WSI
classification pipelines are memory-expensive and computation-inefficient since
they usually assemble tens of thousands of patches as bags for computation. On
the other hand, despite their popularity in other tasks, data augmentations are
unexplored for WSI MIL frameworks. To address them, we propose ReMix, a general
and efficient framework for MIL based WSI classification. It comprises two
steps: reduce and mix. First, it reduces the number of instances in WSI bags by
substituting instances with instance prototypes, i.e., patch cluster centroids.
Then, we propose a ``Mix-the-bag'' augmentation that contains four online,
stochastic and flexible latent space augmentations. It brings diverse and
reliable class-identity-preserving semantic changes in the latent space while
enforcing semantic-perturbation invariance. We evaluate ReMix on two public
datasets with two state-of-the-art MIL methods. In our experiments, consistent
improvements in precision, accuracy, and recall have been achieved but with
orders of magnitude reduced training time and memory consumption, demonstrating
ReMix's effectiveness and efficiency. Code is available.
Related papers
- Breaking the Memory Barrier: Near Infinite Batch Size Scaling for Contrastive Loss [59.835032408496545]
We propose a tile-based strategy that partitions the contrastive loss calculation into arbitrary small blocks.
We also introduce a multi-level tiling strategy to leverage the hierarchical structure of distributed systems.
Compared to SOTA memory-efficient solutions, it achieves a two-order-of-magnitude reduction in memory while maintaining comparable speed.
arXiv Detail & Related papers (2024-10-22T17:59:30Z) - PreMix: Boosting Multiple Instance Learning in Digital Histopathology through Pre-training with Intra-Batch Slide Mixing [2.6703221234079946]
PreMix extends the general MIL framework by pre-training the MIL aggregator with an intra-batch slide mixing approach.
It achieves a mean of 4.7% performance improvement over the baseline MIL framework.
Ultimately, PreMix paves the way for more efficient and accurate WSI classification under limited WSI-labeled datasets.
arXiv Detail & Related papers (2024-08-02T10:24:35Z) - Rethinking Multiple Instance Learning for Whole Slide Image
Classification: A Bag-Level Classifier is a Good Instance-Level Teacher [22.080213609228547]
Multiple Instance Learning has demonstrated promise in Whole Slide Image (WSI) classification.
Existing methods generally adopt a two-stage approach, comprising a non-learnable feature embedding stage and a classifier training stage.
We propose that a bag-level classifier can be a good instance-level teacher.
arXiv Detail & Related papers (2023-12-02T10:16:03Z) - Pseudo-Bag Mixup Augmentation for Multiple Instance Learning-Based Whole
Slide Image Classification [18.679580844360615]
We propose a new Pseudo-bag Mixup (PseMix) data augmentation scheme to improve the training of MIL models.
Our scheme generalizes the Mixup strategy for general images to special WSIs via pseudo-bags.
It is designed as an efficient and decoupled method, neither involving time-consuming operations nor relying on MIL model predictions.
arXiv Detail & Related papers (2023-06-28T13:02:30Z) - Compound Batch Normalization for Long-tailed Image Classification [77.42829178064807]
We propose a compound batch normalization method based on a Gaussian mixture.
It can model the feature space more comprehensively and reduce the dominance of head classes.
The proposed method outperforms existing methods on long-tailed image classification.
arXiv Detail & Related papers (2022-12-02T07:31:39Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - ScoreMix: A Scalable Augmentation Strategy for Training GANs with
Limited Data [93.06336507035486]
Generative Adversarial Networks (GANs) typically suffer from overfitting when limited training data is available.
We present ScoreMix, a novel and scalable data augmentation approach for various image synthesis tasks.
arXiv Detail & Related papers (2022-10-27T02:55:15Z) - Feature Re-calibration based MIL for Whole Slide Image Classification [7.92885032436243]
Whole slide image (WSI) classification is a fundamental task for the diagnosis and treatment of diseases.
We propose to re-calibrate the distribution of a WSI bag (instances) by using the statistics of the max-instance (critical) feature.
We employ a position encoding module (PEM) to model spatial/morphological information, and perform pooling by multi-head self-attention (PSMA) with a Transformer encoder.
arXiv Detail & Related papers (2022-06-22T07:00:39Z) - Improving Calibration for Long-Tailed Recognition [68.32848696795519]
We propose two methods to improve calibration and performance in such scenarios.
For dataset bias due to different samplers, we propose shifted batch normalization.
Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2021-04-01T13:55:21Z) - Attentional-Biased Stochastic Gradient Descent [74.49926199036481]
We present a provable method (named ABSGD) for addressing the data imbalance or label noise problem in deep learning.
Our method is a simple modification to momentum SGD where we assign an individual importance weight to each sample in the mini-batch.
ABSGD is flexible enough to combine with other robust losses without any additional cost.
arXiv Detail & Related papers (2020-12-13T03:41:52Z) - Dual-stream Multiple Instance Learning Network for Whole Slide Image
Classification with Self-supervised Contrastive Learning [16.84711797934138]
We address the challenging problem of whole slide image (WSI) classification.
WSI classification can be cast as a multiple instance learning (MIL) problem when only slide-level labels are available.
We propose a MIL-based method for WSI classification and tumor detection that does not require localized annotations.
arXiv Detail & Related papers (2020-11-17T20:51:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.