LSA: Latent Style Augmentation Towards Stain-Agnostic Cervical Cancer Screening
- URL: http://arxiv.org/abs/2503.06563v1
- Date: Sun, 09 Mar 2025 11:33:59 GMT
- Title: LSA: Latent Style Augmentation Towards Stain-Agnostic Cervical Cancer Screening
- Authors: Jiangdong Cai, Haotian Jiang, Zhenrong Shen, Yonghao Li, Honglin Xiong, Lichi Zhang, Qian Wang,
- Abstract summary: Latent Style Augmentation (LSA) is a framework that performs efficient, online stain augmentation directly on WSI-level latent features.<n>We first introduce WSAug, a WSI-level stain augmentation method ensuring consistent stain across patches within a WSI.<n>Using offline-augmented WSIs by WSAug, we design and train Stain Transformer, which can simulate targeted style in the latent space.
- Score: 12.77650539024463
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The deployment of computer-aided diagnosis systems for cervical cancer screening using whole slide images (WSIs) faces critical challenges due to domain shifts caused by staining variations across different scanners and imaging environments. While existing stain augmentation methods improve patch-level robustness, they fail to scale to WSIs due to two key limitations: (1) inconsistent stain patterns when extending patch operations to gigapixel slides, and (2) prohibitive computational/storage costs from offline processing of augmented WSIs.To address this, we propose Latent Style Augmentation (LSA), a framework that performs efficient, online stain augmentation directly on WSI-level latent features. We first introduce WSAug, a WSI-level stain augmentation method ensuring consistent stain across patches within a WSI. Using offline-augmented WSIs by WSAug, we design and train Stain Transformer, which can simulate targeted style in the latent space, efficiently enhancing the robustness of the WSI-level classifier. We validate our method on a multi-scanner WSI dataset for cervical cancer diagnosis. Despite being trained on data from a single scanner, our approach achieves significant performance improvements on out-of-distribution data from other scanners. Code will be available at https://github.com/caijd2000/LSA.
Related papers
- Vision Transformers with Autoencoders and Explainable AI for Cancer Patient Risk Stratification Using Whole Slide Imaging [3.6940298700319065]
PATH-X is a framework that integrates Vision Transformers (ViT) and Autoencoders with SHAP (Shapley Additive Explanations) to enhance modelability for patient stratification and risk prediction.
A representative image slice is selected from each WSI, and numerical feature embeddings are extracted using Google's pre-trained ViT.
Kaplan-Meier survival analysis is applied to evaluate stratification into two and three risk groups.
arXiv Detail & Related papers (2025-04-07T05:48:42Z) - A self-supervised framework for learning whole slide representations [52.774822784847565]
We present Slide Pre-trained Transformers (SPT) for gigapixel-scale self-supervision of whole slide images.
We benchmark SPT visual representations on five diagnostic tasks across three biomedical microscopy datasets.
arXiv Detail & Related papers (2024-02-09T05:05:28Z) - Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - Artifact-Robust Graph-Based Learning in Digital Pathology [2.9998889086656586]
Whole slide images(WSIs) are digitized images of tissues placed in glass slides using advanced scanners.
In this work, a novel robust learning approach to account for these artifacts is presented.
The accuracy and scores of the proposed model with prostate cancer dataset compared with non-robust algorithms show significant improvement in cancer diagnosis.
arXiv Detail & Related papers (2023-10-27T15:06:01Z) - LESS: Label-efficient Multi-scale Learning for Cytological Whole Slide
Image Screening [19.803614403803962]
We propose a weakly-supervised Label-Efficient WSI Screening method, dubbed LESS, for cytological WSI analysis with only slide-level labels.
We provide appropriate supervision by using slide-level labels to improve the learning of patch-level features.
It outperforms state-of-the-art MIL methods on pathology WSIs and realizes automatic cytological WSI cancer screening.
arXiv Detail & Related papers (2023-06-06T05:09:20Z) - Hierarchical Transformer for Survival Prediction Using Multimodality
Whole Slide Images and Genomics [63.76637479503006]
Learning good representation of giga-pixel level whole slide pathology images (WSI) for downstream tasks is critical.
This paper proposes a hierarchical-based multimodal transformer framework that learns a hierarchical mapping between pathology images and corresponding genes.
Our architecture requires fewer GPU resources compared with benchmark methods while maintaining better WSI representation ability.
arXiv Detail & Related papers (2022-11-29T23:47:56Z) - Patient-level Microsatellite Stability Assessment from Whole Slide
Images By Combining Momentum Contrast Learning and Group Patch Embeddings [6.40476282000118]
Current approaches bypass the WSI high resolution by first classifying small patches extracted from the WSI.
We introduce an effective approach to leverage WSI high resolution information by momentum contrastive learning of patch embeddings.
Our approach achieves up to 7.4% better accuracy compared to the straightforward patch-level classification and patient level aggregation approach.
arXiv Detail & Related papers (2022-08-22T16:31:43Z) - Coarse-to-Fine Sparse Transformer for Hyperspectral Image Reconstruction [138.04956118993934]
We propose a novel Transformer-based method, coarse-to-fine sparse Transformer (CST)
CST embedding HSI sparsity into deep learning for HSI reconstruction.
In particular, CST uses our proposed spectra-aware screening mechanism (SASM) for coarse patch selecting. Then the selected patches are fed into our customized spectra-aggregation hashing multi-head self-attention (SAH-MSA) for fine pixel clustering and self-similarity capturing.
arXiv Detail & Related papers (2022-03-09T16:17:47Z) - DSNet: A Dual-Stream Framework for Weakly-Supervised Gigapixel Pathology
Image Analysis [78.78181964748144]
We present a novel weakly-supervised framework for classifying whole slide images (WSIs)
WSIs are commonly processed by patch-wise classification with patch-level labels.
With image-level labels only, patch-wise classification would be sub-optimal due to inconsistency between the patch appearance and image-level label.
arXiv Detail & Related papers (2021-09-13T09:10:43Z) - An Efficient Cervical Whole Slide Image Analysis Framework Based on
Multi-scale Semantic and Spatial Features using Deep Learning [2.7218168309244652]
This study designs a novel inline connection network (InCNet) by enriching the multi-scale connectivity to build the lightweight model named You Only Look Cytopathology Once (YOLCO)
The proposed model allows the input size enlarged to megapixel that can stitch the WSI without any overlap by the average repeats.
Based on Transformer for classifying the integrated multi-scale multi-task features, the experimental results appear $0.872$ AUC score better and $2.51times$ faster than the best conventional method in WSI classification.
arXiv Detail & Related papers (2021-06-29T06:24:55Z) - A Simple Semi-Supervised Learning Framework for Object Detection [55.95789931533665]
Semi-supervised learning (SSL) has a potential to improve the predictive performance of machine learning models using unlabeled data.
We propose STAC, a simple yet effective SSL framework for visual object detection along with a data augmentation strategy.
arXiv Detail & Related papers (2020-05-10T19:15:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.