Accounting for Dependencies in Deep Learning Based Multiple Instance
Learning for Whole Slide Imaging
- URL: http://arxiv.org/abs/2111.01556v1
- Date: Mon, 1 Nov 2021 06:50:33 GMT
- Title: Accounting for Dependencies in Deep Learning Based Multiple Instance
Learning for Whole Slide Imaging
- Authors: Andriy Myronenko, Ziyue Xu, Dong Yang, Holger Roth, Daguang Xu
- Abstract summary: Multiple instance learning (MIL) is a key algorithm for classification of whole slide images (WSI)
Histology WSIs can have billions of pixels, which create enormous computational and annotation challenges.
We propose an instance-wise loss function based on instance pseudo-labels.
- Score: 8.712556146101953
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multiple instance learning (MIL) is a key algorithm for classification of
whole slide images (WSI). Histology WSIs can have billions of pixels, which
create enormous computational and annotation challenges. Typically, such images
are divided into a set of patches (a bag of instances), where only bag-level
class labels are provided. Deep learning based MIL methods calculate instance
features using convolutional neural network (CNN). Our proposed approach is
also deep learning based, with the following two contributions: Firstly, we
propose to explicitly account for dependencies between instances during
training by embedding self-attention Transformer blocks to capture dependencies
between instances. For example, a tumor grade may depend on the presence of
several particular patterns at different locations in WSI, which requires to
account for dependencies between patches. Secondly, we propose an instance-wise
loss function based on instance pseudo-labels. We compare the proposed
algorithm to multiple baseline methods, evaluate it on the PANDA challenge
dataset, the largest publicly available WSI dataset with over 11K images, and
demonstrate state-of-the-art results.
Related papers
- MamMIL: Multiple Instance Learning for Whole Slide Images with State Space Models [56.37780601189795]
We propose a framework named MamMIL for WSI analysis.
We represent each WSI as an undirected graph.
To address the problem that Mamba can only process 1D sequences, we propose a topology-aware scanning mechanism.
arXiv Detail & Related papers (2024-03-08T09:02:13Z) - RoFormer for Position Aware Multiple Instance Learning in Whole Slide
Image Classification [0.0]
Whole slide image (WSI) classification is a critical task in computational pathology.
Current methods rely on multiple-instance learning (MIL) models with frozen feature extractors.
We show that our method outperforms state-of-the-art MIL models on weakly supervised classification tasks.
arXiv Detail & Related papers (2023-10-03T09:59:59Z) - Self-similarity Driven Scale-invariant Learning for Weakly Supervised
Person Search [66.95134080902717]
We propose a novel one-step framework, named Self-similarity driven Scale-invariant Learning (SSL)
We introduce a Multi-scale Exemplar Branch to guide the network in concentrating on the foreground and learning scale-invariant features.
Experiments on PRW and CUHK-SYSU databases demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2023-02-25T04:48:11Z) - Hierarchical Transformer for Survival Prediction Using Multimodality
Whole Slide Images and Genomics [63.76637479503006]
Learning good representation of giga-pixel level whole slide pathology images (WSI) for downstream tasks is critical.
This paper proposes a hierarchical-based multimodal transformer framework that learns a hierarchical mapping between pathology images and corresponding genes.
Our architecture requires fewer GPU resources compared with benchmark methods while maintaining better WSI representation ability.
arXiv Detail & Related papers (2022-11-29T23:47:56Z) - Multiple Instance Learning via Iterative Self-Paced Supervised
Contrastive Learning [22.07044031105496]
Learning representations for individual instances when only bag-level labels are available is a challenge in multiple instance learning (MIL)
We propose a novel framework, Iterative Self-paced Supervised Contrastive Learning for MIL Representations (ItS2CLR)
It improves the learned representation by exploiting instance-level pseudo labels derived from the bag-level labels.
arXiv Detail & Related papers (2022-10-17T21:43:32Z) - Boosting Few-shot Fine-grained Recognition with Background Suppression
and Foreground Alignment [53.401889855278704]
Few-shot fine-grained recognition (FS-FGR) aims to recognize novel fine-grained categories with the help of limited available samples.
We propose a two-stage background suppression and foreground alignment framework, which is composed of a background activation suppression (BAS) module, a foreground object alignment (FOA) module, and a local to local (L2L) similarity metric.
Experiments conducted on multiple popular fine-grained benchmarks demonstrate that our method outperforms the existing state-of-the-art by a large margin.
arXiv Detail & Related papers (2022-10-04T07:54:40Z) - ReMix: A General and Efficient Framework for Multiple Instance Learning
based Whole Slide Image Classification [14.78430890440035]
Whole slide image (WSI) classification often relies on weakly supervised multiple instance learning (MIL) methods to handle gigapixel resolution images and slide-level labels.
We propose ReMix, a general and efficient framework for MIL based WSI classification.
arXiv Detail & Related papers (2022-07-05T04:21:35Z) - Multi-level Second-order Few-shot Learning [111.0648869396828]
We propose a Multi-level Second-order (MlSo) few-shot learning network for supervised or unsupervised few-shot image classification and few-shot action recognition.
We leverage so-called power-normalized second-order base learner streams combined with features that express multiple levels of visual abstraction.
We demonstrate respectable results on standard datasets such as Omniglot, mini-ImageNet, tiered-ImageNet, Open MIC, fine-grained datasets such as CUB Birds, Stanford Dogs and Cars, and action recognition datasets such as HMDB51, UCF101, and mini-MIT.
arXiv Detail & Related papers (2022-01-15T19:49:00Z) - Deep Relational Metric Learning [84.95793654872399]
This paper presents a deep relational metric learning framework for image clustering and retrieval.
We learn an ensemble of features that characterizes an image from different aspects to model both interclass and intraclass distributions.
Experiments on the widely-used CUB-200-2011, Cars196, and Stanford Online Products datasets demonstrate that our framework improves existing deep metric learning methods and achieves very competitive results.
arXiv Detail & Related papers (2021-08-23T09:31:18Z) - Sparse convolutional context-aware multiple instance learning for whole
slide image classification [7.18791111462057]
Whole slide microscopic slides display many cues about the underlying tissue guiding diagnostic and the choice of therapy for many diseases.
To tackle this issue, multiple instance learning (MIL) classifies bags of patches instead of whole slide images.
Our approach presents a paradigm shift through the integration of spatial information of patches with a sparse-input convolutional-based MIL strategy.
arXiv Detail & Related papers (2021-05-06T14:46:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.