Rethinking Multiple Instance Learning for Whole Slide Image
Classification: A Bag-Level Classifier is a Good Instance-Level Teacher
- URL: http://arxiv.org/abs/2312.01099v1
- Date: Sat, 2 Dec 2023 10:16:03 GMT
- Title: Rethinking Multiple Instance Learning for Whole Slide Image
Classification: A Bag-Level Classifier is a Good Instance-Level Teacher
- Authors: Hongyi Wang, Luyang Luo, Fang Wang, Ruofeng Tong, Yen-Wei Chen,
Hongjie Hu, Lanfen Lin, Hao Chen
- Abstract summary: Multiple Instance Learning has demonstrated promise in Whole Slide Image (WSI) classification.
Existing methods generally adopt a two-stage approach, comprising a non-learnable feature embedding stage and a classifier training stage.
We propose that a bag-level classifier can be a good instance-level teacher.
- Score: 22.080213609228547
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multiple Instance Learning (MIL) has demonstrated promise in Whole Slide
Image (WSI) classification. However, a major challenge persists due to the high
computational cost associated with processing these gigapixel images. Existing
methods generally adopt a two-stage approach, comprising a non-learnable
feature embedding stage and a classifier training stage. Though it can greatly
reduce the memory consumption by using a fixed feature embedder pre-trained on
other domains, such scheme also results in a disparity between the two stages,
leading to suboptimal classification accuracy. To address this issue, we
propose that a bag-level classifier can be a good instance-level teacher. Based
on this idea, we design Iteratively Coupled Multiple Instance Learning (ICMIL)
to couple the embedder and the bag classifier at a low cost. ICMIL initially
fix the patch embedder to train the bag classifier, followed by fixing the bag
classifier to fine-tune the patch embedder. The refined embedder can then
generate better representations in return, leading to a more accurate
classifier for the next iteration. To realize more flexible and more effective
embedder fine-tuning, we also introduce a teacher-student framework to
efficiently distill the category knowledge in the bag classifier to help the
instance-level embedder fine-tuning. Thorough experiments were conducted on
four distinct datasets to validate the effectiveness of ICMIL. The experimental
results consistently demonstrate that our method significantly improves the
performance of existing MIL backbones, achieving state-of-the-art results. The
code is available at: https://github.com/Dootmaan/ICMIL/tree/confidence_based
Related papers
- Rethinking Multiple Instance Learning for Whole Slide Image Classification: A Good Instance Classifier is All You Need [18.832471712088353]
We propose an instance-level weakly supervised contrastive learning algorithm for the first time under the MIL setting.
We also propose an accurate pseudo label generation method through prototype learning.
arXiv Detail & Related papers (2023-07-05T12:44:52Z) - Iteratively Coupled Multiple Instance Learning from Instance to Bag
Classifier for Whole Slide Image Classification [21.16848269555692]
Whole Slide Image (WSI) classification remains a challenge due to their extremely high resolution and the absence of fine-grained labels.
We propose a novel framework called Iteratively Coupled MIL (ICMIL) which bridges the loss back-propagation process from the bag-level classifier to the patch embedder.
arXiv Detail & Related papers (2023-03-28T06:12:53Z) - Learning Context-aware Classifier for Semantic Segmentation [88.88198210948426]
In this paper, contextual hints are exploited via learning a context-aware classifier.
Our method is model-agnostic and can be easily applied to generic segmentation models.
With only negligible additional parameters and +2% inference time, decent performance gain has been achieved on both small and large models.
arXiv Detail & Related papers (2023-03-21T07:00:35Z) - Prediction Calibration for Generalized Few-shot Semantic Segmentation [101.69940565204816]
Generalized Few-shot Semantic (GFSS) aims to segment each image pixel into either base classes with abundant training examples or novel classes with only a handful of (e.g., 1-5) training images per class.
We build a cross-attention module that guides the classifier's final prediction using the fused multi-level features.
Our PCN outperforms the state-the-art alternatives by large margins.
arXiv Detail & Related papers (2022-10-15T13:30:12Z) - Bi-directional Weakly Supervised Knowledge Distillation for Whole Slide
Image Classification [9.43604501642743]
We propose an end-to-end weakly supervised knowledge distillation framework (WENO) for WSI classification.
In this paper, we propose an end-to-end weakly supervised knowledge distillation framework (WENO) for WSI classification.
arXiv Detail & Related papers (2022-10-07T16:12:04Z) - ReMix: A General and Efficient Framework for Multiple Instance Learning
based Whole Slide Image Classification [14.78430890440035]
Whole slide image (WSI) classification often relies on weakly supervised multiple instance learning (MIL) methods to handle gigapixel resolution images and slide-level labels.
We propose ReMix, a general and efficient framework for MIL based WSI classification.
arXiv Detail & Related papers (2022-07-05T04:21:35Z) - No Fear of Heterogeneity: Classifier Calibration for Federated Learning
with Non-IID Data [78.69828864672978]
A central challenge in training classification models in the real-world federated system is learning with non-IID data.
We propose a novel and simple algorithm called Virtual Representations (CCVR), which adjusts the classifier using virtual representations sampled from an approximated ssian mixture model.
Experimental results demonstrate that CCVR state-of-the-art performance on popular federated learning benchmarks including CIFAR-10, CIFAR-100, and CINIC-10.
arXiv Detail & Related papers (2021-06-09T12:02:29Z) - Improving Calibration for Long-Tailed Recognition [68.32848696795519]
We propose two methods to improve calibration and performance in such scenarios.
For dataset bias due to different samplers, we propose shifted batch normalization.
Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2021-04-01T13:55:21Z) - Prior Guided Feature Enrichment Network for Few-Shot Segmentation [64.91560451900125]
State-of-the-art semantic segmentation methods require sufficient labeled data to achieve good results.
Few-shot segmentation is proposed to tackle this problem by learning a model that quickly adapts to new classes with a few labeled support samples.
Theses frameworks still face the challenge of generalization ability reduction on unseen classes due to inappropriate use of high-level semantic information.
arXiv Detail & Related papers (2020-08-04T10:41:32Z) - Dual-stream Maximum Self-attention Multi-instance Learning [11.685285490589981]
Multi-instance learning (MIL) is a form of weakly supervised learning where a single class label is assigned to a bag of instances while the instance-level labels are not available.
We propose a dual-stream maximum self-attention MIL model (DSMIL) parameterized by neural networks.
Our method achieves superior performance compared to the best MIL methods and demonstrates state-of-the-art performance on benchmark MIL datasets.
arXiv Detail & Related papers (2020-06-09T22:44:58Z) - SCAN: Learning to Classify Images without Labels [73.69513783788622]
We advocate a two-step approach where feature learning and clustering are decoupled.
A self-supervised task from representation learning is employed to obtain semantically meaningful features.
We obtain promising results on ImageNet, and outperform several semi-supervised learning methods in the low-data regime.
arXiv Detail & Related papers (2020-05-25T18:12:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.