Task-specific Fine-tuning via Variational Information Bottleneck for
Weakly-supervised Pathology Whole Slide Image Classification
- URL: http://arxiv.org/abs/2303.08446v1
- Date: Wed, 15 Mar 2023 08:41:57 GMT
- Title: Task-specific Fine-tuning via Variational Information Bottleneck for
Weakly-supervised Pathology Whole Slide Image Classification
- Authors: Honglin Li, Chenglu Zhu, Yunlong Zhang, Yuxuan Sun, Zhongyi Shui,
Wenwei Kuang, Sunyi Zheng, Lin Yang
- Abstract summary: Multiple Instance Learning (MIL) has shown promising results in digital Pathology Whole Slide Image (WSI) classification.
We propose an efficient WSI fine-tuning framework motivated by the Information Bottleneck theory.
Our framework is evaluated on five pathology WSI datasets on various WSI heads.
- Score: 10.243293283318415
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While Multiple Instance Learning (MIL) has shown promising results in digital
Pathology Whole Slide Image (WSI) classification, such a paradigm still faces
performance and generalization problems due to challenges in high computational
costs on Gigapixel WSIs and limited sample size for model training. To deal
with the computation problem, most MIL methods utilize a frozen pretrained
model from ImageNet to obtain representations first. This process may lose
essential information owing to the large domain gap and hinder the
generalization of model due to the lack of image-level training-time
augmentations. Though Self-supervised Learning (SSL) proposes viable
representation learning schemes, the improvement of the downstream task still
needs to be further explored in the conversion from the task-agnostic features
of SSL to the task-specifics under the partial label supervised learning. To
alleviate the dilemma of computation cost and performance, we propose an
efficient WSI fine-tuning framework motivated by the Information Bottleneck
theory. The theory enables the framework to find the minimal sufficient
statistics of WSI, thus supporting us to fine-tune the backbone into a
task-specific representation only depending on WSI-level weak labels. The
WSI-MIL problem is further analyzed to theoretically deduce our fine-tuning
method. Our framework is evaluated on five pathology WSI datasets on various
WSI heads. The experimental results of our fine-tuned representations show
significant improvements in both accuracy and generalization compared with
previous works. Source code will be available at
https://github.com/invoker-LL/WSI-finetuning.
Related papers
- MergeUp-augmented Semi-Weakly Supervised Learning for WSI Classification [1.2387547097768696]
Multiple instance learning (MIL) is a promising weakly supervised learning approach for WSI classification.
We introduce a feature augmentation technique, MergeUp, which merges bags with low-priority bags to enhance inter-category information.
Experimental results on the CAMELYON-16, BRACS, and TCGA-LUNG datasets demonstrate the superiority of our method over existing state-of-the-art approaches.
arXiv Detail & Related papers (2024-08-23T04:08:30Z) - MSCPT: Few-shot Whole Slide Image Classification with Multi-scale and Context-focused Prompt Tuning [11.717352903130411]
Multiple instance learning (MIL) has become a standard paradigm for weakly supervised classification of whole slide images (WSI)
The lack of training data and the presence of rare diseases present significant challenges for these methods.
We propose a Multi-Scale and Context-focused Prompt Tuning (MSCPT) method for FSWC tasks.
arXiv Detail & Related papers (2024-08-21T10:25:51Z) - MamMIL: Multiple Instance Learning for Whole Slide Images with State Space Models [56.37780601189795]
We propose a framework named MamMIL for WSI analysis.
We represent each WSI as an undirected graph.
To address the problem that Mamba can only process 1D sequences, we propose a topology-aware scanning mechanism.
arXiv Detail & Related papers (2024-03-08T09:02:13Z) - Generalizable Whole Slide Image Classification with Fine-Grained Visual-Semantic Interaction [17.989559761931435]
We propose a novel "Fine-grained Visual-Semantic Interaction" framework for WSI classification.
It is designed to enhance the model's generalizability by leveraging the interaction between localized visual patterns and fine-grained pathological semantics.
Our method demonstrates robust generalizability and strong transferability, dominantly outperforming the counterparts on the TCGA Lung Cancer dataset.
arXiv Detail & Related papers (2024-02-29T16:29:53Z) - Tumor segmentation on whole slide images: training or prompting? [0.0]
We show the efficacy of visual prompting in the context of tumor segmentation for three distinct organs.
Our findings reveal that, with appropriate prompt examples, visual prompting can achieve comparable or better performance without extensive fine-tuning.
arXiv Detail & Related papers (2024-02-21T16:59:53Z) - A self-supervised framework for learning whole slide representations [52.774822784847565]
We present Slide Pre-trained Transformers (SPT) for gigapixel-scale self-supervision of whole slide images.
We benchmark SPT visual representations on five diagnostic tasks across three biomedical microscopy datasets.
arXiv Detail & Related papers (2024-02-09T05:05:28Z) - A General Framework for Learning from Weak Supervision [93.89870459388185]
This paper introduces a general framework for learning from weak supervision (GLWS) with a novel algorithm.
Central to GLWS is an Expectation-Maximization (EM) formulation, adeptly accommodating various weak supervision sources.
We also present an advanced algorithm that significantly simplifies the EM computational demands.
arXiv Detail & Related papers (2024-02-02T21:48:50Z) - BROW: Better featuRes fOr Whole slide image based on self-distillation [19.295596638166536]
Whole slide image (WSI) processing is becoming part of the key components of standard clinical diagnosis for various diseases.
The performance of most WSI-related tasks relies on the efficacy of the backbone which extracts WSI patch feature representations.
We proposed BROW, a foundation model for extracting better feature representations for WSIs, which can be conveniently adapted to downstream tasks without or with slight fine-tuning.
arXiv Detail & Related papers (2023-09-15T09:11:09Z) - Hierarchical Transformer for Survival Prediction Using Multimodality
Whole Slide Images and Genomics [63.76637479503006]
Learning good representation of giga-pixel level whole slide pathology images (WSI) for downstream tasks is critical.
This paper proposes a hierarchical-based multimodal transformer framework that learns a hierarchical mapping between pathology images and corresponding genes.
Our architecture requires fewer GPU resources compared with benchmark methods while maintaining better WSI representation ability.
arXiv Detail & Related papers (2022-11-29T23:47:56Z) - SUPERB: Speech processing Universal PERformance Benchmark [78.41287216481203]
Self-supervised learning (SSL) has proven vital for advancing research in natural language processing (NLP) and computer vision (CV)
SuperB is a leaderboard to benchmark the performance of a shared model across a wide range of speech processing tasks.
We present a simple framework to solve SUPERB tasks by learning task-specialized lightweight prediction heads on top of the frozen shared model.
arXiv Detail & Related papers (2021-05-03T17:51:09Z) - Prior Guided Feature Enrichment Network for Few-Shot Segmentation [64.91560451900125]
State-of-the-art semantic segmentation methods require sufficient labeled data to achieve good results.
Few-shot segmentation is proposed to tackle this problem by learning a model that quickly adapts to new classes with a few labeled support samples.
Theses frameworks still face the challenge of generalization ability reduction on unseen classes due to inappropriate use of high-level semantic information.
arXiv Detail & Related papers (2020-08-04T10:41:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.