RoFormer for Position Aware Multiple Instance Learning in Whole Slide
Image Classification
- URL: http://arxiv.org/abs/2310.01924v1
- Date: Tue, 3 Oct 2023 09:59:59 GMT
- Title: RoFormer for Position Aware Multiple Instance Learning in Whole Slide
Image Classification
- Authors: Etienne Pochet, Rami Maroun, Roger Trullo
- Abstract summary: Whole slide image (WSI) classification is a critical task in computational pathology.
Current methods rely on multiple-instance learning (MIL) models with frozen feature extractors.
We show that our method outperforms state-of-the-art MIL models on weakly supervised classification tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Whole slide image (WSI) classification is a critical task in computational
pathology. However, the gigapixel-size of such images remains a major challenge
for the current state of deep-learning. Current methods rely on
multiple-instance learning (MIL) models with frozen feature extractors. Given
the the high number of instances in each image, MIL methods have long assumed
independence and permutation-invariance of patches, disregarding the tissue
structure and correlation between patches. Recent works started studying this
correlation between instances but the computational workload of such a high
number of tokens remained a limiting factor. In particular, relative position
of patches remains unaddressed. We propose to apply a straightforward encoding
module, namely a RoFormer layer , relying on memory-efficient exact
self-attention and relative positional encoding. This module can perform full
self-attention with relative position encoding on patches of large and
arbitrary shaped WSIs, solving the need for correlation between instances and
spatial modeling of tissues. We demonstrate that our method outperforms
state-of-the-art MIL models on three commonly used public datasets (TCGA-NSCLC,
BRACS and Camelyon16)) on weakly supervised classification tasks. Code is
available at https://github.com/Sanofi-Public/DDS-RoFormerMIL
Related papers
- MamMIL: Multiple Instance Learning for Whole Slide Images with State Space Models [56.37780601189795]
We propose a framework named MamMIL for WSI analysis.
We represent each WSI as an undirected graph.
To address the problem that Mamba can only process 1D sequences, we propose a topology-aware scanning mechanism.
arXiv Detail & Related papers (2024-03-08T09:02:13Z) - Learned representation-guided diffusion models for large-image generation [58.192263311786824]
We introduce a novel approach that trains diffusion models conditioned on embeddings from self-supervised learning (SSL)
Our diffusion models successfully project these features back to high-quality histopathology and remote sensing images.
Augmenting real data by generating variations of real images improves downstream accuracy for patch-level and larger, image-scale classification tasks.
arXiv Detail & Related papers (2023-12-12T14:45:45Z) - Deep Multiple Instance Learning with Distance-Aware Self-Attention [9.361964965928063]
We introduce a novel multiple instance learning (MIL) model with distance-aware self-attention (DAS-MIL)
Unlike existing relative position representations for self-attention which are discrete, our approach introduces continuous distance-dependent terms into the computation of the attention weights.
We evaluate our model on a custom MNIST-based MIL dataset and on CAMELYON16, a publicly available cancer metastasis detection dataset.
arXiv Detail & Related papers (2023-05-17T20:11:43Z) - Hierarchical Transformer for Survival Prediction Using Multimodality
Whole Slide Images and Genomics [63.76637479503006]
Learning good representation of giga-pixel level whole slide pathology images (WSI) for downstream tasks is critical.
This paper proposes a hierarchical-based multimodal transformer framework that learns a hierarchical mapping between pathology images and corresponding genes.
Our architecture requires fewer GPU resources compared with benchmark methods while maintaining better WSI representation ability.
arXiv Detail & Related papers (2022-11-29T23:47:56Z) - Gigapixel Whole-Slide Images Classification using Locally Supervised
Learning [31.213316201151954]
Histo whole slide images (WSIs) play a very important role in clinical studies and serve as the gold standard for many cancer diagnoses.
Conventional methods rely on a multiple instance learning (MIL) strategy to process a WSI at patch level.
We propose a locally supervised learning framework which processes the entire slide by exploring the entire local and global information.
arXiv Detail & Related papers (2022-07-17T19:31:54Z) - Feature Re-calibration based MIL for Whole Slide Image Classification [7.92885032436243]
Whole slide image (WSI) classification is a fundamental task for the diagnosis and treatment of diseases.
We propose to re-calibrate the distribution of a WSI bag (instances) by using the statistics of the max-instance (critical) feature.
We employ a position encoding module (PEM) to model spatial/morphological information, and perform pooling by multi-head self-attention (PSMA) with a Transformer encoder.
arXiv Detail & Related papers (2022-06-22T07:00:39Z) - Decoupled Multi-task Learning with Cyclical Self-Regulation for Face
Parsing [71.19528222206088]
We propose a novel Decoupled Multi-task Learning with Cyclical Self-Regulation for face parsing.
Specifically, DML-CSR designs a multi-task model which comprises face parsing, binary edge, and category edge detection.
Our method achieves the new state-of-the-art performance on the Helen, CelebA-HQ, and LapaMask datasets.
arXiv Detail & Related papers (2022-03-28T02:12:30Z) - Coarse-to-Fine Sparse Transformer for Hyperspectral Image Reconstruction [138.04956118993934]
We propose a novel Transformer-based method, coarse-to-fine sparse Transformer (CST)
CST embedding HSI sparsity into deep learning for HSI reconstruction.
In particular, CST uses our proposed spectra-aware screening mechanism (SASM) for coarse patch selecting. Then the selected patches are fed into our customized spectra-aggregation hashing multi-head self-attention (SAH-MSA) for fine pixel clustering and self-similarity capturing.
arXiv Detail & Related papers (2022-03-09T16:17:47Z) - Accounting for Dependencies in Deep Learning Based Multiple Instance
Learning for Whole Slide Imaging [8.712556146101953]
Multiple instance learning (MIL) is a key algorithm for classification of whole slide images (WSI)
Histology WSIs can have billions of pixels, which create enormous computational and annotation challenges.
We propose an instance-wise loss function based on instance pseudo-labels.
arXiv Detail & Related papers (2021-11-01T06:50:33Z) - Sparse convolutional context-aware multiple instance learning for whole
slide image classification [7.18791111462057]
Whole slide microscopic slides display many cues about the underlying tissue guiding diagnostic and the choice of therapy for many diseases.
To tackle this issue, multiple instance learning (MIL) classifies bags of patches instead of whole slide images.
Our approach presents a paradigm shift through the integration of spatial information of patches with a sparse-input convolutional-based MIL strategy.
arXiv Detail & Related papers (2021-05-06T14:46:09Z) - Cross-Scale Internal Graph Neural Network for Image Super-Resolution [147.77050877373674]
Non-local self-similarity in natural images has been well studied as an effective prior in image restoration.
For single image super-resolution (SISR), most existing deep non-local methods only exploit similar patches within the same scale of the low-resolution (LR) input image.
This is achieved using a novel cross-scale internal graph neural network (IGNN)
arXiv Detail & Related papers (2020-06-30T10:48:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.