Revisiting Data Challenges of Computational Pathology: A Pack-based Multiple Instance Learning Framework
- URL: http://arxiv.org/abs/2509.20923v1
- Date: Thu, 25 Sep 2025 09:05:40 GMT
- Title: Revisiting Data Challenges of Computational Pathology: A Pack-based Multiple Instance Learning Framework
- Authors: Wenhao Tang, Heng Fang, Ge Wu, Xiang Li, Ming-Ming Cheng,
- Abstract summary: Computational pathology (CPath) digitizes pathology slides into whole slide images (WSIs)<n>WSIs possess extremely long sequence lengths (up to 200K), significant length variations (from 200 to 200K), and limited supervision.<n>We propose a pack-based MIL framework to address these challenges.
- Score: 47.035885218675126
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Computational pathology (CPath) digitizes pathology slides into whole slide images (WSIs), enabling analysis for critical healthcare tasks such as cancer diagnosis and prognosis. However, WSIs possess extremely long sequence lengths (up to 200K), significant length variations (from 200 to 200K), and limited supervision. These extreme variations in sequence length lead to high data heterogeneity and redundancy. Conventional methods often compromise on training efficiency and optimization to preserve such heterogeneity under limited supervision. To comprehensively address these challenges, we propose a pack-based MIL framework. It packs multiple sampled, variable-length feature sequences into fixed-length ones, enabling batched training while preserving data heterogeneity. Moreover, we introduce a residual branch that composes discarded features from multiple slides into a hyperslide which is trained with tailored labels. It offers multi-slide supervision while mitigating feature loss from sampling. Meanwhile, an attention-driven downsampler is introduced to compress features in both branches to reduce redundancy. By alleviating these challenges, our approach achieves an accuracy improvement of up to 8% while using only 12% of the training time in the PANDA(UNI). Extensive experiments demonstrate that focusing data challenges in CPath holds significant potential in the era of foundation models. The code is https://github.com/FangHeng/PackMIL
Related papers
- MambaMIL+: Modeling Long-Term Contextual Patterns for Gigapixel Whole Slide Image [24.093388981091717]
Multiple instance learning (MIL) offers a solution by treating each WSI as a bag of patch-level instances.<n>Mamba has emerged as a promising alternative for long sequence learning, scaling linearly to thousands of tokens.<n>We propose MambaMIL+, a new MIL framework that explicitly integrates spatial context while maintaining long-range dependency modeling.
arXiv Detail & Related papers (2025-12-19T16:01:14Z) - Unsupervised Online 3D Instance Segmentation with Synthetic Sequences and Dynamic Loss [52.28880405119483]
Unsupervised online 3D instance segmentation is a fundamental yet challenging task.<n>Existing methods, such as UNIT, have made progress in this direction but remain constrained by limited training diversity.<n>We propose a new framework that enriches the training distribution through synthetic point cloud sequence generation.
arXiv Detail & Related papers (2025-09-27T08:53:27Z) - PHEATPRUNER: Interpretable Data-centric Feature Selection for Multivariate Time Series Classification through Persistent Homology [0.0937465283958018]
PHeatPruner is a method to balance performance and interpretability in time series classification.<n> Persistent homology facilitates the pruning of up to 45% of the applied variables.<n>Sheaf theory contributes explanatory vectors that provide deeper insights into the data's structural nuances.
arXiv Detail & Related papers (2025-04-25T13:14:11Z) - Reshaping the Online Data Buffering and Organizing Mechanism for Continual Test-Time Adaptation [49.53202761595912]
Continual Test-Time Adaptation involves adapting a pre-trained source model to continually changing unsupervised target domains.
We analyze the challenges of this task: online environment, unsupervised nature, and the risks of error accumulation and catastrophic forgetting.
We propose an uncertainty-aware buffering approach to identify and aggregate significant samples with high certainty from the unsupervised, single-pass data stream.
arXiv Detail & Related papers (2024-07-12T15:48:40Z) - TSLANet: Rethinking Transformers for Time Series Representation Learning [19.795353886621715]
Time series data is characterized by its intrinsic long and short-range dependencies.
We introduce a novel Time Series Lightweight Network (TSLANet) as a universal convolutional model for diverse time series tasks.
Our experiments demonstrate that TSLANet outperforms state-of-the-art models in various tasks spanning classification, forecasting, and anomaly detection.
arXiv Detail & Related papers (2024-04-12T13:41:29Z) - Primal Dual Continual Learning: Balancing Stability and Plasticity through Adaptive Memory Allocation [86.8475564814154]
We show that it is both possible and beneficial to undertake the constrained optimization problem directly.
We focus on memory-based methods, where a small subset of samples from previous tasks can be stored in a replay buffer.
We show that dual variables indicate the sensitivity of the optimal value of the continual learning problem with respect to constraint perturbations.
arXiv Detail & Related papers (2023-09-29T21:23:27Z) - MUSTANG: Multi-Stain Self-Attention Graph Multiple Instance Learning
Pipeline for Histopathology Whole Slide Images [1.127806343149511]
Whole Slide Images (WSIs) present a challenging computer vision task due to their gigapixel size and presence of artefacts.
Real-world clinical datasets tend to come as sets of heterogeneous WSIs with labels present at the patient-level, with poor to no annotations.
Here we propose an end-to-end multi-stain self-attention graph (MUSTANG) multiple instance learning pipeline.
arXiv Detail & Related papers (2023-09-19T14:30:14Z) - ScoreMix: A Scalable Augmentation Strategy for Training GANs with
Limited Data [93.06336507035486]
Generative Adversarial Networks (GANs) typically suffer from overfitting when limited training data is available.
We present ScoreMix, a novel and scalable data augmentation approach for various image synthesis tasks.
arXiv Detail & Related papers (2022-10-27T02:55:15Z) - Mixed-Precision Neural Network Quantization via Learned Layer-wise
Importance [50.00102219630088]
Mixed-precision quantization (MPQ) makes it hard to determine the optimal bit-width for each layer.
We propose a joint training scheme that can obtain all indicators at once.
For example, MPQ search on ResNet18 with our indicators takes only 0.06 seconds.
arXiv Detail & Related papers (2022-03-16T03:23:50Z) - Pattern Discovery in Time Series with Byte Pair Encoding [12.338599136651261]
We propose an unsupervised method for learning representations of time series based on common patterns identified within them.
In this way the method can capture both long-term and short-term dependencies present in the data.
arXiv Detail & Related papers (2021-05-30T00:47:19Z) - Brain Image Synthesis with Unsupervised Multivariate Canonical
CSC$\ell_4$Net [122.8907826672382]
We propose to learn dedicated features that cross both intre- and intra-modal variations using a novel CSC$ell_4$Net.
arXiv Detail & Related papers (2021-03-22T05:19:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.