Information-driven Fusion of Pathology Foundation Models for Enhanced Disease Characterization
- URL: http://arxiv.org/abs/2512.11104v1
- Date: Thu, 11 Dec 2025 20:38:03 GMT
- Title: Information-driven Fusion of Pathology Foundation Models for Enhanced Disease Characterization
- Authors: Brennan Flannery, Thomas DeSilvio, Jane Nguyen, Satish E. Viswanath,
- Abstract summary: Foundation models (FMs) have demonstrated strong performance across diverse pathology tasks.<n>We propose an information-driven, intelligent fusion strategy for integrating multiple FMs into a unified representation.<n>Our findings suggest that intelligent, correlation-guided fusion of pathology FMs can yield compact, task-tailored representations.
- Score: 0.26249027950824505
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Foundation models (FMs) have demonstrated strong performance across diverse pathology tasks. While there are similarities in the pre-training objectives of FMs, there is still limited understanding of their complementarity, redundancy in embedding spaces, or biological interpretation of features. In this study, we propose an information-driven, intelligent fusion strategy for integrating multiple pathology FMs into a unified representation and systematically evaluate its performance for cancer grading and staging across three distinct diseases. Diagnostic H&E whole-slide images from kidney (519 slides), prostate (490 slides), and rectal (200 slides) cancers were dichotomized into low versus high grade or stage. Both tile-level FMs (Conch v1.5, MUSK, Virchow2, H-Optimus1, Prov-Gigapath) and slide-level FMs (TITAN, CHIEF, MADELEINE) were considered to train downstream classifiers. We then evaluated three FM fusion schemes at both tile and slide levels: majority-vote ensembling, naive feature concatenation, and intelligent fusion based on correlation-guided pruning of redundant features. Under patient-stratified cross-validation with hold-out testing, intelligent fusion of tile-level embeddings yielded consistent gains in classification performance across all three cancers compared with the best single FMs and naive fusion. Global similarity metrics revealed substantial alignment of FM embedding spaces, contrasted by lower local neighborhood agreement, indicating complementary fine-grained information across FMs. Attention maps showed that intelligent fusion yielded concentrated attention on tumor regions while reducing spurious focus on benign regions. Our findings suggest that intelligent, correlation-guided fusion of pathology FMs can yield compact, task-tailored representations that enhance both predictive performance and interpretability in downstream computational pathology tasks.
Related papers
- Multimodal Visual Surrogate Compression for Alzheimer's Disease Classification [69.87877580725768]
Multimodal Visual Surrogate Compression (MVSC) learns to compress and adapt large 3D sMRI volumes into compact 2D features.<n>MVSC has two key components: a Volume Context that captures global cross-slice context under textual guidance, and an Adaptive Slice Fusion module that aggregates slice-level information in a text-enhanced, patch-wise manner.
arXiv Detail & Related papers (2026-01-29T13:05:46Z) - Proof of Concept for Mammography Classification with Enhanced Compactness and Separability Modules [0.0]
This study presents a validation and extension of a recent methodological framework for medical image classification.<n>Using a Kaggle dataset that consolidates INbreast, MIAS, and InceptionM mammography collections, we compare a baseline CNN, ConvNeXt Tiny, and InceptionV3 backbones enriched with GAGM and SE modules.<n>Results confirm the effectiveness of GAGM and SE in enhancing feature discriminability and reducing false negatives.<n>In our experiments, however, the Feature Smoothing Loss did not yield measurable improvements under mammography classification conditions.
arXiv Detail & Related papers (2025-12-06T21:36:05Z) - Collaborative Attention and Consistent-Guided Fusion of MRI and PET for Alzheimer's Disease Diagnosis [12.33741976057116]
Alzheimer's disease (AD) is the most prevalent form of dementia, and its early diagnosis is essential for slowing disease progression.<n>Recent studies on multimodal neuroimaging fusion using MRI and PET have achieved promising results.<n>We propose a Collaborative Attention and Consistent-Guided Fusion framework for MRI and PET based AD diagnosis.
arXiv Detail & Related papers (2025-11-04T03:42:07Z) - Fusion of Heterogeneous Pathology Foundation Models for Whole Slide Image Analysis [10.323462166785133]
Whole slide image (WSI) analysis has emerged as an increasingly essential technique in computational pathology.<n>Recent advances in pathological foundation models (FMs) have demonstrated significant advantages in deriving meaningful patch-level or slide-level feature representations from WSIs.<n>We propose a novel framework for the fusion of heterogeneous pathological FMs, called FuseCPath, yielding a model with a superior ensemble performance.
arXiv Detail & Related papers (2025-10-31T06:59:11Z) - MedSeqFT: Sequential Fine-tuning Foundation Models for 3D Medical Image Segmentation [55.37355146924576]
MedSeqFT is a sequential fine-tuning framework for medical image analysis.<n>It adapts pre-trained models to new tasks while refining their representational capacity.<n>It consistently outperforms state-of-the-art fine-tuning strategies.
arXiv Detail & Related papers (2025-09-07T15:22:53Z) - FusionFM: Fusing Eye-specific Foundational Models for Optimized Ophthalmic Diagnosis [36.79693801937608]
Foundation models (FMs) have shown great promise in medical image analysis by improving generalization across diverse downstream tasks.<n>To our knowledge, this is the first study to systematically evaluate both single and fused ophthalmic FMs.<n>We benchmarked four state-of-the-art FMs using standardized datasets from multiple countries and evaluated their performance using AUC and F1 metrics.
arXiv Detail & Related papers (2025-08-15T01:17:52Z) - AdaFusion: Prompt-Guided Inference with Adaptive Fusion of Pathology Foundation Models [49.550545038402184]
We propose AdaFusion, a novel prompt-guided inference framework.<n>Our method compresses and aligns tile-level features from diverse models.<n>AdaFusion consistently surpasses individual PFMs across both classification and regression tasks.
arXiv Detail & Related papers (2025-08-07T07:09:31Z) - MIRROR: Multi-Modal Pathological Self-Supervised Representation Learning via Modality Alignment and Retention [57.044719143401664]
Histopathology and transcriptomics are fundamental modalities in oncology, encapsulating the morphological and molecular aspects of the disease.<n>We present MIRROR, a novel multi-modal representation learning method designed to foster both modality alignment and retention.<n>Extensive evaluations on TCGA cohorts for cancer subtyping and survival analysis highlight MIRROR's superior performance.
arXiv Detail & Related papers (2025-03-01T07:02:30Z) - Multimodal Outer Arithmetic Block Dual Fusion of Whole Slide Images and Omics Data for Precision Oncology [6.418265127069878]
We propose the use of omic embeddings during early and late fusion to capture complementary information from local (patch-level) to global (slide-level) interactions.<n>This dual fusion strategy enhances interpretability and classification performance, highlighting its potential for clinical diagnostics.
arXiv Detail & Related papers (2024-11-26T13:25:53Z) - An Interpretable Cross-Attentive Multi-modal MRI Fusion Framework for Schizophrenia Diagnosis [46.58592655409785]
We propose a novel Cross-Attentive Multi-modal Fusion framework (CAMF) to capture both intra-modal and inter-modal relationships between fMRI and sMRI.
Our approach significantly improves classification accuracy, as demonstrated by our evaluations on two extensive multi-modal brain imaging datasets.
The gradient-guided Score-CAM is applied to interpret critical functional networks and brain regions involved in schizophrenia.
arXiv Detail & Related papers (2024-03-29T20:32:30Z) - Brain Imaging-to-Graph Generation using Adversarial Hierarchical Diffusion Models for MCI Causality Analysis [44.45598796591008]
Brain imaging-to-graph generation (BIGG) framework is proposed to map functional magnetic resonance imaging (fMRI) into effective connectivity for mild cognitive impairment analysis.
The hierarchical transformers in the generator are designed to estimate the noise at multiple scales.
Evaluations of the ADNI dataset demonstrate the feasibility and efficacy of the proposed model.
arXiv Detail & Related papers (2023-05-18T06:54:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.