Prompt-Free SAM-Based Multi-Task Framework for Breast Ultrasound Lesion Segmentation and Classification
- URL: http://arxiv.org/abs/2601.05498v1
- Date: Fri, 09 Jan 2026 03:02:41 GMT
- Title: Prompt-Free SAM-Based Multi-Task Framework for Breast Ultrasound Lesion Segmentation and Classification
- Authors: Samuel E. Johnny, Bernes L. Atabonfack, Israel Alagbe, Assane Gueye,
- Abstract summary: This study presents a multi-task deep learning framework that jointly performs lesion segmentation and diagnostic classification.<n>Our approach employs a prompt-free, fully supervised adaptation where high-dimensional SAM features are decoded through either a lightweight convolutional head or a UNet-inspired decoder for pixel-wise segmentation.<n> Experiments on the PRECISE 2025 breast ultrasound dataset, split per class into 80 percent training and 20 percent testing, show that the proposed method achieves a Dice Similarity Coefficient (DSC) of 0.887 and an accuracy of 92.3 percent.
- Score: 0.4083182125683813
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Accurate tumor segmentation and classification in breast ultrasound (BUS) imaging remain challenging due to low contrast, speckle noise, and diverse lesion morphology. This study presents a multi-task deep learning framework that jointly performs lesion segmentation and diagnostic classification using embeddings from the Segment Anything Model (SAM) vision encoder. Unlike prompt-based SAM variants, our approach employs a prompt-free, fully supervised adaptation where high-dimensional SAM features are decoded through either a lightweight convolutional head or a UNet-inspired decoder for pixel-wise segmentation. The classification branch is enhanced via mask-guided attention, allowing the model to focus on lesion-relevant features while suppressing background artifacts. Experiments on the PRECISE 2025 breast ultrasound dataset, split per class into 80 percent training and 20 percent testing, show that the proposed method achieves a Dice Similarity Coefficient (DSC) of 0.887 and an accuracy of 92.3 percent, ranking among the top entries on the PRECISE challenge leaderboard. These results demonstrate that SAM-based representations, when coupled with segmentation-guided learning, significantly improve both lesion delineation and diagnostic prediction in breast ultrasound imaging.
Related papers
- Hide-and-Seek Attribution: Weakly Supervised Segmentation of Vertebral Metastases in CT [68.09387763135236]
We introduce a weakly supervised method trained solely on vertebra-level healthy/malignant labels, without any lesion masks.<n>We achieve strong blastic/lytic performance despite no mask supervision.
arXiv Detail & Related papers (2025-12-07T14:03:28Z) - Transformer Classification of Breast Lesions: The BreastDCEDL_AMBL Benchmark Dataset and 0.92 AUC Baseline [1.9336815376402718]
This study introduces a transformer-based framework for automated classification of breast lesions in dynamic contrast-enhanced MRI.<n>We implemented a SegFormer architecture that achieved an AUC of 0.92 for lesion-level classification, with 100% sensitivity and 67% specificity at the patient level.<n>Public release of the dataset, models, and evaluation protocols provides the first standardized benchmark for DCE-MRI lesion classification.
arXiv Detail & Related papers (2025-09-30T15:58:02Z) - Multimodal Deep Learning for Phyllodes Tumor Classification from Ultrasound and Clinical Data [0.29981448312652675]
Phyllodes tumors (PTs) are difficult to classify preoperatively due to their radiological similarity to benign fibroadenomas.<n>We propose a multimodal deep learning framework that integrates breast ultrasound (BUS) images with structured clinical data to improve diagnostic accuracy.
arXiv Detail & Related papers (2025-08-29T19:54:11Z) - SAM-Driven Weakly Supervised Nodule Segmentation with Uncertainty-Aware Cross Teaching [13.5553526185399]
Automated nodule segmentation is essential for computer-assisted diagnosis in ultrasound images.
Recently, segmentation foundation models like SAM have shown impressive generalizability on natural images.
In this work, we devise a novel weakly supervised framework that effectively utilizes the segmentation foundation model to generate pseudo-labels.
arXiv Detail & Related papers (2024-07-18T14:27:54Z) - Semi- and Weakly-Supervised Learning for Mammogram Mass Segmentation with Limited Annotations [49.33388736227072]
We propose a semi- and weakly-supervised learning framework for mass segmentation.
We use limited strongly-labeled samples and sufficient weakly-labeled samples to achieve satisfactory performance.
Experiments on CBIS-DDSM and INbreast datasets demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2024-03-14T12:05:25Z) - Morphology-Enhanced CAM-Guided SAM for weakly supervised Breast Lesion Segmentation [7.747608350830482]
We present a novel framework for weakly supervised lesion segmentation in early breast ultrasound images.
Our method uses morphological enhancement and class activation map (CAM)-guided localization.
This approach does not require pixel-level annotation, thereby reducing the cost of data annotation.
arXiv Detail & Related papers (2023-11-18T22:06:04Z) - Rethinking Semi-Supervised Medical Image Segmentation: A
Variance-Reduction Perspective [51.70661197256033]
We propose ARCO, a semi-supervised contrastive learning framework with stratified group theory for medical image segmentation.
We first propose building ARCO through the concept of variance-reduced estimation and show that certain variance-reduction techniques are particularly beneficial in pixel/voxel-level segmentation tasks.
We experimentally validate our approaches on eight benchmarks, i.e., five 2D/3D medical and three semantic segmentation datasets, with different label settings.
arXiv Detail & Related papers (2023-02-03T13:50:25Z) - Reliable Joint Segmentation of Retinal Edema Lesions in OCT Images [55.83984261827332]
In this paper, we propose a novel reliable multi-scale wavelet-enhanced transformer network.
We develop a novel segmentation backbone that integrates a wavelet-enhanced feature extractor network and a multi-scale transformer module.
Our proposed method achieves better segmentation accuracy with a high degree of reliability as compared to other state-of-the-art segmentation approaches.
arXiv Detail & Related papers (2022-12-01T07:32:56Z) - Generative Adversarial Networks for Weakly Supervised Generation and Evaluation of Brain Tumor Segmentations on MR Images [0.0]
This work presents a weakly supervised approach to segment anomalies in 2D magnetic resonance images.
We train a generative adversarial network (GAN) that converts cancerous images to healthy variants.
Non-cancerous variants can also be used to evaluate the segmentations in a weakly supervised fashion.
arXiv Detail & Related papers (2022-11-10T00:04:46Z) - Mixed-UNet: Refined Class Activation Mapping for Weakly-Supervised
Semantic Segmentation with Multi-scale Inference [28.409679398886304]
We develop a novel model named Mixed-UNet, which has two parallel branches in the decoding phase.
We evaluate the designed Mixed-UNet against several prevalent deep learning-based segmentation approaches on our dataset collected from the local hospital and public datasets.
arXiv Detail & Related papers (2022-05-06T08:37:02Z) - Malignancy Prediction and Lesion Identification from Clinical
Dermatological Images [65.1629311281062]
We consider machine-learning-based malignancy prediction and lesion identification from clinical dermatological images.
We first identify all lesions present in the image regardless of sub-type or likelihood of malignancy, then it estimates their likelihood of malignancy, and through aggregation, it also generates an image-level likelihood of malignancy.
arXiv Detail & Related papers (2021-04-02T20:52:05Z) - Improved Slice-wise Tumour Detection in Brain MRIs by Computing
Dissimilarities between Latent Representations [68.8204255655161]
Anomaly detection for Magnetic Resonance Images (MRIs) can be solved with unsupervised methods.
We have proposed a slice-wise semi-supervised method for tumour detection based on the computation of a dissimilarity function in the latent space of a Variational AutoEncoder.
We show that by training the models on higher resolution images and by improving the quality of the reconstructions, we obtain results which are comparable with different baselines.
arXiv Detail & Related papers (2020-07-24T14:02:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.