MUSTANG: Multi-Stain Self-Attention Graph Multiple Instance Learning
Pipeline for Histopathology Whole Slide Images
- URL: http://arxiv.org/abs/2309.10650v2
- Date: Wed, 4 Oct 2023 14:24:09 GMT
- Title: MUSTANG: Multi-Stain Self-Attention Graph Multiple Instance Learning
Pipeline for Histopathology Whole Slide Images
- Authors: Amaya Gallagher-Syed, Luca Rossi, Felice Rivellese, Costantino
Pitzalis, Myles Lewis, Michael Barnes, Gregory Slabaugh
- Abstract summary: Whole Slide Images (WSIs) present a challenging computer vision task due to their gigapixel size and presence of artefacts.
Real-world clinical datasets tend to come as sets of heterogeneous WSIs with labels present at the patient-level, with poor to no annotations.
Here we propose an end-to-end multi-stain self-attention graph (MUSTANG) multiple instance learning pipeline.
- Score: 1.127806343149511
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Whole Slide Images (WSIs) present a challenging computer vision task due to
their gigapixel size and presence of numerous artefacts. Yet they are a
valuable resource for patient diagnosis and stratification, often representing
the gold standard for diagnostic tasks. Real-world clinical datasets tend to
come as sets of heterogeneous WSIs with labels present at the patient-level,
with poor to no annotations. Weakly supervised attention-based multiple
instance learning approaches have been developed in recent years to address
these challenges, but can fail to resolve both long and short-range
dependencies. Here we propose an end-to-end multi-stain self-attention graph
(MUSTANG) multiple instance learning pipeline, which is designed to solve a
weakly-supervised gigapixel multi-image classification task, where the label is
assigned at the patient-level, but no slide-level labels or region annotations
are available. The pipeline uses a self-attention based approach by restricting
the operations to a highly sparse k-Nearest Neighbour Graph of embedded WSI
patches based on the Euclidean distance. We show this approach achieves a
state-of-the-art F1-score/AUC of 0.89/0.92, outperforming the widely used CLAM
model. Our approach is highly modular and can easily be modified to suit
different clinical datasets, as it only requires a patient-level label without
annotations and accepts WSI sets of different sizes, as the graphs can be of
varying sizes and structures. The source code can be found at
https://github.com/AmayaGS/MUSTANG.
Related papers
- Semantic Segmentation Based Quality Control of Histopathology Whole Slide Images [2.953447779233234]
We developed a software pipeline for quality control (QC) of histopathology whole slide images (WSIs)
It segments various regions, such as blurs of different levels, tissue regions, tissue folds, and pen marks.
It was evaluated in all TCGAs, which is the largest publicly available WSI dataset containing more than 11,000 histopathology images from 28 organs.
arXiv Detail & Related papers (2024-10-04T10:03:04Z) - MamMIL: Multiple Instance Learning for Whole Slide Images with State Space Models [56.37780601189795]
We propose a framework named MamMIL for WSI analysis.
We represent each WSI as an undirected graph.
To address the problem that Mamba can only process 1D sequences, we propose a topology-aware scanning mechanism.
arXiv Detail & Related papers (2024-03-08T09:02:13Z) - Long-MIL: Scaling Long Contextual Multiple Instance Learning for
Histopathology Whole Slide Image Analysis [9.912061800841267]
Whole Slide Image (WSI) of histopathology tissue is used for analysis.
Previous methods generally divide the WSI into a large number of patches, then aggregate all patches within a WSI to make the slide-level prediction.
We propose to amend position embedding for shape varying long-contextual WSI by introducing Linear Bias into Attention.
arXiv Detail & Related papers (2023-11-21T03:08:47Z) - Context-Aware Self-Supervised Learning of Whole Slide Images [0.0]
A novel two-stage learning technique is presented in this work.
A graph representation capturing all dependencies among regions in the WSI is very intuitive.
The entire slide is presented as a graph, where the nodes correspond to the patches from the WSI.
The proposed framework is then tested using WSIs from prostate and kidney cancers.
arXiv Detail & Related papers (2023-06-07T20:23:05Z) - Self-similarity Driven Scale-invariant Learning for Weakly Supervised
Person Search [66.95134080902717]
We propose a novel one-step framework, named Self-similarity driven Scale-invariant Learning (SSL)
We introduce a Multi-scale Exemplar Branch to guide the network in concentrating on the foreground and learning scale-invariant features.
Experiments on PRW and CUHK-SYSU databases demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2023-02-25T04:48:11Z) - Hierarchical Transformer for Survival Prediction Using Multimodality
Whole Slide Images and Genomics [63.76637479503006]
Learning good representation of giga-pixel level whole slide pathology images (WSI) for downstream tasks is critical.
This paper proposes a hierarchical-based multimodal transformer framework that learns a hierarchical mapping between pathology images and corresponding genes.
Our architecture requires fewer GPU resources compared with benchmark methods while maintaining better WSI representation ability.
arXiv Detail & Related papers (2022-11-29T23:47:56Z) - A graph-transformer for whole slide image classification [11.968797693846476]
We present a Graph-Transformer (GT) that fuses a graph-based representation of an whole slide image (WSI) and a vision transformer for processing pathology images, called GTP, to predict disease grade.
Our findings demonstrate GTP as an interpretable and effective deep learning framework for WSI-level classification.
arXiv Detail & Related papers (2022-05-19T16:32:10Z) - Spatial-spectral Hyperspectral Image Classification via Multiple Random
Anchor Graphs Ensemble Learning [88.60285937702304]
This paper proposes a novel spatial-spectral HSI classification method via multiple random anchor graphs ensemble learning (RAGE)
Firstly, the local binary pattern is adopted to extract the more descriptive features on each selected band, which preserves local structures and subtle changes of a region.
Secondly, the adaptive neighbors assignment is introduced in the construction of anchor graph, to reduce the computational complexity.
arXiv Detail & Related papers (2021-03-25T09:31:41Z) - Cluster-to-Conquer: A Framework for End-to-End Multi-Instance Learning
for Whole Slide Image Classification [7.876654642325896]
We propose an end-to-end framework that clusters the patches from a Whole Slide Images (WSI) into $k$-groups, samples $k'$ patches from each group for training, and uses an adaptive attention mechanism for slide level prediction.
The framework is optimized end-to-end on slide-level cross-entropy, patch-level cross-entropy, and KL-divergence loss.
arXiv Detail & Related papers (2021-03-19T04:24:01Z) - Attention-Driven Dynamic Graph Convolutional Network for Multi-Label
Image Recognition [53.17837649440601]
We propose an Attention-Driven Dynamic Graph Convolutional Network (ADD-GCN) to dynamically generate a specific graph for each image.
Experiments on public multi-label benchmarks demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2020-12-05T10:10:12Z) - Knowledge-Guided Multi-Label Few-Shot Learning for General Image
Recognition [75.44233392355711]
KGGR framework exploits prior knowledge of statistical label correlations with deep neural networks.
It first builds a structured knowledge graph to correlate different labels based on statistical label co-occurrence.
Then, it introduces the label semantics to guide learning semantic-specific features.
It exploits a graph propagation network to explore graph node interactions.
arXiv Detail & Related papers (2020-09-20T15:05:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.