SG-CLDFF: A Novel Framework for Automated White Blood Cell Classification and Segmentation
- URL: http://arxiv.org/abs/2510.17278v1
- Date: Mon, 20 Oct 2025 08:07:39 GMT
- Title: SG-CLDFF: A Novel Framework for Automated White Blood Cell Classification and Segmentation
- Authors: Mehdi Zekriyapanah Gashti, Mostafa Mohammadpour, Ghasem Farjamnia,
- Abstract summary: Saliency-Guided Cross-Layer Deep Feature Fusion framework (SG-CLDFF)<n>A lightweight hybrid backbone (Swin-style) produces multi-resolution representations, which are fused by a ResNeXt-CCinspired cross-layer fusion module.<n>Interpretability is enforced through Grad-CAM visualizations and saliency consistency checks, allowing model decisions to be inspected at the regional level.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Accurate segmentation and classification of white blood cells (WBCs) in microscopic images are essential for diagnosis and monitoring of many hematological disorders, yet remain challenging due to staining variability, complex backgrounds, and class imbalance. In this paper, we introduce a novel Saliency-Guided Cross-Layer Deep Feature Fusion framework (SG-CLDFF) that tightly integrates saliency-driven preprocessing with multi-scale deep feature aggregation to improve both robustness and interpretability for WBC analysis. SG-CLDFF first computes saliency priors to highlight candidate WBC regions and guide subsequent feature extraction. A lightweight hybrid backbone (EfficientSwin-style) produces multi-resolution representations, which are fused by a ResNeXt-CC-inspired cross-layer fusion module to preserve complementary information from shallow and deep layers. The network is trained in a multi-task setup with concurrent segmentation and cell-type classification heads, using class-aware weighted losses and saliency-alignment regularization to mitigate imbalance and suppress background activation. Interpretability is enforced through Grad-CAM visualizations and saliency consistency checks, allowing model decisions to be inspected at the regional level. We validate the framework on standard public benchmarks (BCCD, LISC, ALL-IDB), reporting consistent gains in IoU, F1, and classification accuracy compared to strong CNN and transformer baselines. An ablation study also demonstrates the individual contributions of saliency preprocessing and cross-layer fusion. SG-CLDFF offers a practical and explainable path toward more reliable automated WBC analysis in clinical workflows.
Related papers
- scMRDR: A scalable and flexible framework for unpaired single-cell multi-omics data integration [53.683726781791385]
We introduce a scalable and flexible generative framework called single-cell Multi-omics Regularized Disentangled Representations (scMRDR) for unpaired multi-omics integration.<n>Our method achieves excellent performance on benchmark datasets in terms of batch correction, modality alignment, and biological signal preservation.
arXiv Detail & Related papers (2025-10-28T21:28:39Z) - EndoCIL: A Class-Incremental Learning Framework for Endoscopic Image Classification [5.574295682041076]
Class-incremental learning (CIL) for endoscopic image analysis is crucial for real-world clinical applications.<n>We propose EndoCIL, a novel and unified CIL framework specifically tailored for endoscopic image diagnosis.
arXiv Detail & Related papers (2025-10-20T06:26:54Z) - Comparative Analysis of Data Augmentation for Clinical ECG Classification with STAR [0.0]
Sinusoidal Time--Amplitude Resampling (STAR) is a beat-wise augmentation that operates strictly between successive R-peaks.<n>STAR is designed for practical pipelines and offers: (i) morphology-faithful variability that broadens training diversity without corrupting peaks or intervals; (ii) source-resilient training, improving stability across devices, sites, and cohorts without dataset-specific tuning; and (iv) better learning on rare classes via beat-level augmentation.
arXiv Detail & Related papers (2025-10-15T14:18:03Z) - Enhanced SegNet with Integrated Grad-CAM for Interpretable Retinal Layer Segmentation in OCT Images [0.0]
This study proposes an improved SegNet-based deep learning framework for automated and interpretable retinal layer segmentation.<n> Architectural innovations, including modified pooling strategies, enhance feature extraction from noisy OCT images.<n>Grad-CAM visualizations highlighted anatomically relevant regions, aligning segmentation with clinical biomarkers.
arXiv Detail & Related papers (2025-09-09T14:31:51Z) - Uncertainty-aware Cross-training for Semi-supervised Medical Image Segmentation [45.96892342675963]
We propose an Uncertainty-aware Cross-training framework for semi-supervised medical image (UC-Seg)<n>Our method achieves superior segmentation accuracy and generalization performance compared to other state-of-the-art semi-supervised methods.
arXiv Detail & Related papers (2025-08-12T15:28:10Z) - ActiveSSF: An Active-Learning-Guided Self-Supervised Framework for Long-Tailed Megakaryocyte Classification [3.6535793744942318]
We propose the ActiveSSF framework, which integrates active learning with self-supervised pretraining.<n> Experimental results on clinical megakaryocyte datasets demonstrate that ActiveSSF achieves state-of-the-art performance.
arXiv Detail & Related papers (2025-02-12T08:24:36Z) - KaLDeX: Kalman Filter based Linear Deformable Cross Attention for Retina Vessel Segmentation [46.57880203321858]
We propose a novel network (KaLDeX) for vascular segmentation leveraging a Kalman filter based linear deformable cross attention (LDCA) module.
Our approach is based on two key components: Kalman filter (KF) based linear deformable convolution (LD) and cross-attention (CA) modules.
The proposed method is evaluated on retinal fundus image datasets (DRIVE, CHASE_BD1, and STARE) as well as the 3mm and 6mm of the OCTA-500 dataset.
arXiv Detail & Related papers (2024-10-28T16:00:42Z) - DAFFNet: A Dual Attention Feature Fusion Network for Classification of White Blood Cells [2.0005570775461567]
We propose a novel dual-branch network Dual Attention Feature Fusion Network (DAFFNet), which integrates the high-level semantic features with morphological features of WBC.
Our proposed network framework achieves 98.77%, 91.30%, 98.36%, 99.71%, 98.45%, and 98.85% overall accuracy on the six public datasets.
arXiv Detail & Related papers (2024-05-25T13:09:25Z) - ConDiSR: Contrastive Disentanglement and Style Regularization for Single Domain Generalization [42.810247034149214]
Medical data often exhibits distribution shifts, which cause test-time performance degradation for deep learning models trained using standard pipelines.
This study highlights the importance and challenges of exploring Single Domain Generalization frameworks in the context of the classification task.
arXiv Detail & Related papers (2024-03-14T13:50:44Z) - Reliable Joint Segmentation of Retinal Edema Lesions in OCT Images [55.83984261827332]
In this paper, we propose a novel reliable multi-scale wavelet-enhanced transformer network.
We develop a novel segmentation backbone that integrates a wavelet-enhanced feature extractor network and a multi-scale transformer module.
Our proposed method achieves better segmentation accuracy with a high degree of reliability as compared to other state-of-the-art segmentation approaches.
arXiv Detail & Related papers (2022-12-01T07:32:56Z) - Co-Heterogeneous and Adaptive Segmentation from Multi-Source and
Multi-Phase CT Imaging Data: A Study on Pathological Liver and Lesion
Segmentation [48.504790189796836]
We present a novel segmentation strategy, co-heterogenous and adaptive segmentation (CHASe)
We propose a versatile framework that fuses appearance based semi-supervision, mask based adversarial domain adaptation, and pseudo-labeling.
CHASe can further improve pathological liver mask Dice-Sorensen coefficients by ranges of $4.2% sim 9.4%$.
arXiv Detail & Related papers (2020-05-27T06:58:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.