Feature Re-calibration based MIL for Whole Slide Image Classification
- URL: http://arxiv.org/abs/2206.10878v1
- Date: Wed, 22 Jun 2022 07:00:39 GMT
- Title: Feature Re-calibration based MIL for Whole Slide Image Classification
- Authors: Philip Chikontwe, Soo Jeong Nam, Heounjeong Go, Meejeong Kim, Hyun
Jung Sung, Sang Hyun Park
- Abstract summary: Whole slide image (WSI) classification is a fundamental task for the diagnosis and treatment of diseases.
We propose to re-calibrate the distribution of a WSI bag (instances) by using the statistics of the max-instance (critical) feature.
We employ a position encoding module (PEM) to model spatial/morphological information, and perform pooling by multi-head self-attention (PSMA) with a Transformer encoder.
- Score: 7.92885032436243
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Whole slide image (WSI) classification is a fundamental task for the
diagnosis and treatment of diseases; but, curation of accurate labels is
time-consuming and limits the application of fully-supervised methods. To
address this, multiple instance learning (MIL) is a popular method that poses
classification as a weakly supervised learning task with slide-level labels
only. While current MIL methods apply variants of the attention mechanism to
re-weight instance features with stronger models, scant attention is paid to
the properties of the data distribution. In this work, we propose to
re-calibrate the distribution of a WSI bag (instances) by using the statistics
of the max-instance (critical) feature. We assume that in binary MIL, positive
bags have larger feature magnitudes than negatives, thus we can enforce the
model to maximize the discrepancy between bags with a metric feature loss that
models positive bags as out-of-distribution. To achieve this, unlike existing
MIL methods that use single-batch training modes, we propose balanced-batch
sampling to effectively use the feature loss i.e., (+/-) bags simultaneously.
Further, we employ a position encoding module (PEM) to model
spatial/morphological information, and perform pooling by multi-head
self-attention (PSMA) with a Transformer encoder. Experimental results on
existing benchmark datasets show our approach is effective and improves over
state-of-the-art MIL methods.
Related papers
- Attention Is Not What You Need: Revisiting Multi-Instance Learning for Whole Slide Image Classification [51.95824566163554]
We argue that synergizing the standard MIL assumption with variational inference encourages the model to focus on tumour morphology instead of spurious correlations.
Our method also achieves better classification boundaries for identifying hard instances and mitigates the effect of spurious correlations between bags and labels.
arXiv Detail & Related papers (2024-08-18T12:15:22Z) - Bring Metric Functions into Diffusion Models [145.71911023514252]
We introduce a Cascaded Diffusion Model (Cas-DM) that improves a Denoising Diffusion Probabilistic Model (DDPM)
The proposed diffusion model backbone enables the effective use of the LPIPS loss, leading to state-of-the-art image quality (FID, sFID, IS)
Experiment results show that the proposed diffusion model backbone enables the effective use of the LPIPS loss, leading to state-of-the-art image quality (FID, sFID, IS)
arXiv Detail & Related papers (2024-01-04T18:55:01Z) - Self-Supervised Neuron Segmentation with Multi-Agent Reinforcement
Learning [53.00683059396803]
Mask image model (MIM) has been widely used due to its simplicity and effectiveness in recovering original information from masked images.
We propose a decision-based MIM that utilizes reinforcement learning (RL) to automatically search for optimal image masking ratio and masking strategy.
Our approach has a significant advantage over alternative self-supervised methods on the task of neuron segmentation.
arXiv Detail & Related papers (2023-10-06T10:40:46Z) - Interventional Bag Multi-Instance Learning On Whole-Slide Pathological
Images [30.82201485481452]
Bag contextual prior may trick the model into capturing spurious correlations between bags and labels.
We propose a novel scheme, Interventional Bag Multi-Instance Learning (IBMIL), to achieve deconfounded bag-level prediction.
arXiv Detail & Related papers (2023-03-13T05:46:56Z) - DGMIL: Distribution Guided Multiple Instance Learning for Whole Slide
Image Classification [9.950131528559211]
We propose a feature distribution guided deep MIL framework for WSI classification and positive patch localization.
Experiments on the CAMELYON16 dataset and the TCGA Lung Cancer dataset show that our method achieves new SOTA for both global classification and positive patch localization tasks.
arXiv Detail & Related papers (2022-06-17T16:04:30Z) - Momentum Pseudo-Labeling for Semi-Supervised Speech Recognition [55.362258027878966]
We present momentum pseudo-labeling (MPL) as a simple yet effective strategy for semi-supervised speech recognition.
MPL consists of a pair of online and offline models that interact and learn from each other, inspired by the mean teacher method.
The experimental results demonstrate that MPL effectively improves over the base model and is scalable to different semi-supervised scenarios.
arXiv Detail & Related papers (2021-06-16T16:24:55Z) - Dual-stream Multiple Instance Learning Network for Whole Slide Image
Classification with Self-supervised Contrastive Learning [16.84711797934138]
We address the challenging problem of whole slide image (WSI) classification.
WSI classification can be cast as a multiple instance learning (MIL) problem when only slide-level labels are available.
We propose a MIL-based method for WSI classification and tumor detection that does not require localized annotations.
arXiv Detail & Related papers (2020-11-17T20:51:15Z) - Sparse Network Inversion for Key Instance Detection in Multiple Instance
Learning [24.66638752977373]
Multiple Instance Learning (MIL) involves predicting a single label for a bag of instances, given positive or negative labels at bag-level.
The attention-based deep MIL model is a recent advance in both bag-level classification and key instance detection.
We present a method to improve the attention-based deep MIL model in the task of KID.
arXiv Detail & Related papers (2020-09-07T07:01:59Z) - Multi-Scale Positive Sample Refinement for Few-Shot Object Detection [61.60255654558682]
Few-shot object detection (FSOD) helps detectors adapt to unseen classes with few training instances.
We propose a Multi-scale Positive Sample Refinement (MPSR) approach to enrich object scales in FSOD.
MPSR generates multi-scale positive samples as object pyramids and refines the prediction at various scales.
arXiv Detail & Related papers (2020-07-18T09:48:29Z) - Dual-stream Maximum Self-attention Multi-instance Learning [11.685285490589981]
Multi-instance learning (MIL) is a form of weakly supervised learning where a single class label is assigned to a bag of instances while the instance-level labels are not available.
We propose a dual-stream maximum self-attention MIL model (DSMIL) parameterized by neural networks.
Our method achieves superior performance compared to the best MIL methods and demonstrates state-of-the-art performance on benchmark MIL datasets.
arXiv Detail & Related papers (2020-06-09T22:44:58Z) - Weakly-Supervised Action Localization with Expectation-Maximization
Multi-Instance Learning [82.41415008107502]
Weakly-supervised action localization requires training a model to localize the action segments in the video given only video level action label.
It can be solved under the Multiple Instance Learning (MIL) framework, where a bag (video) contains multiple instances (action segments)
We show that our EM-MIL approach more accurately models both the learning objective and the MIL assumptions.
arXiv Detail & Related papers (2020-03-31T23:36:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.