Auto-Generating Weak Labels for Real & Synthetic Data to Improve Label-Scarce Medical Image Segmentation
- URL: http://arxiv.org/abs/2404.17033v1
- Date: Thu, 25 Apr 2024 20:47:08 GMT
- Title: Auto-Generating Weak Labels for Real & Synthetic Data to Improve Label-Scarce Medical Image Segmentation
- Authors: Tanvi Deshpande, Eva Prakash, Elsie Gyang Ross, Curtis Langlotz, Andrew Ng, Jeya Maria Jose Valanarasu,
- Abstract summary: We present a new approach to overcome the hurdle of costly medical image labeling by leveraging foundation models like Segment Anything Model (SAM) and its medical alternate MedSAM.
Our pipeline has the ability to generate weak labels for any unlabeled medical image and subsequently use it to augment label-scarce datasets.
This automation eliminates the manual prompting step in MedSAM, creating a streamlined process for generating labels for both real and synthetic images.
- Score: 5.368714143438489
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The high cost of creating pixel-by-pixel gold-standard labels, limited expert availability, and presence of diverse tasks make it challenging to generate segmentation labels to train deep learning models for medical imaging tasks. In this work, we present a new approach to overcome the hurdle of costly medical image labeling by leveraging foundation models like Segment Anything Model (SAM) and its medical alternate MedSAM. Our pipeline has the ability to generate weak labels for any unlabeled medical image and subsequently use it to augment label-scarce datasets. We perform this by leveraging a model trained on a few gold-standard labels and using it to intelligently prompt MedSAM for weak label generation. This automation eliminates the manual prompting step in MedSAM, creating a streamlined process for generating labels for both real and synthetic images, regardless of quantity. We conduct experiments on label-scarce settings for multiple tasks pertaining to modalities ranging from ultrasound, dermatology, and X-rays to demonstrate the usefulness of our pipeline. The code is available at https://github.com/stanfordmlgroup/Auto-Generate-WLs/.
Related papers
- Self-Relaxed Joint Training: Sample Selection for Severity Estimation with Ordinal Noisy Labels [5.892066196730197]
We propose a new framework for training with ordinal'' noisy labels.
Our framework uses two techniques: clean sample selection and dual-network architecture.
By appropriately using the soft and hard labels in the two techniques, we achieve more accurate sample selection and robust network training.
arXiv Detail & Related papers (2024-10-29T09:23:09Z) - Medical Image Segmentation with SAM-generated Annotations [12.432602118806573]
We evaluate the performance of the Segment Anything Model (SAM) as an annotation tool for medical data.
We generate so-called "pseudo labels" on the Medical Decathlon (MSD) computed tomography (CT) tasks.
The pseudo labels are then used in place of ground truth labels to train a UNet model in a weakly-supervised manner.
arXiv Detail & Related papers (2024-09-30T12:43:20Z) - Deep Self-Cleansing for Medical Image Segmentation with Noisy Labels [33.676420623855314]
Medical image segmentation is crucial in the field of medical imaging, aiding in disease diagnosis and surgical planning.
Most established segmentation methods rely on supervised deep learning, in which clean and precise labels are essential for supervision.
We propose a deep self-cleansing segmentation framework that can preserve clean labels while cleansing noisy ones in the training phase.
arXiv Detail & Related papers (2024-09-08T08:33:32Z) - Automatic Labels are as Effective as Manual Labels in Biomedical Images Classification with Deep Learning [32.71326861917332]
One of the main limitations to train deep learning algorithms to perform a specific task is the need for medical experts to label data.
This paper aims to investigate under which circumstances automatic labels can be adopted to train a DL model on the classification of Whole Slide Images (WSI)
The application of the Semantic Knowledge Extractor Tool (SKET) algorithm to generate automatic labels leads to performance comparable to the one obtained with manual labels.
arXiv Detail & Related papers (2024-06-20T14:20:50Z) - Distilling Self-Supervised Vision Transformers for Weakly-Supervised
Few-Shot Classification & Segmentation [58.03255076119459]
We address the task of weakly-supervised few-shot image classification and segmentation, by leveraging a Vision Transformer (ViT)
Our proposed method takes token representations from the self-supervised ViT and leverages their correlations, via self-attention, to produce classification and segmentation predictions.
Experiments on Pascal-5i and COCO-20i demonstrate significant performance gains in a variety of supervision settings.
arXiv Detail & Related papers (2023-07-07T06:16:43Z) - Losses over Labels: Weakly Supervised Learning via Direct Loss
Construction [71.11337906077483]
Programmable weak supervision is a growing paradigm within machine learning.
We propose Losses over Labels (LoL) as it creates losses directly from ofs without going through the intermediate step of a label.
We show that LoL improves upon existing weak supervision methods on several benchmark text and image classification tasks.
arXiv Detail & Related papers (2022-12-13T22:29:14Z) - BoMD: Bag of Multi-label Descriptors for Noisy Chest X-ray
Classification [25.76256302330625]
New medical imaging classification problems may need to rely on machine-generated noisy labels extracted from radiology reports.
Current noisy-label learning methods designed for multi-class problems cannot be easily adapted.
We propose a new method designed for the noisy multi-label CXR learning, which detects and smoothly re-labels samples from the dataset.
arXiv Detail & Related papers (2022-03-03T08:04:59Z) - Structured Semantic Transfer for Multi-Label Recognition with Partial
Labels [85.6967666661044]
We propose a structured semantic transfer (SST) framework that enables training multi-label recognition models with partial labels.
The framework consists of two complementary transfer modules that explore within-image and cross-image semantic correlations.
Experiments on the Microsoft COCO, Visual Genome and Pascal VOC datasets show that the proposed SST framework obtains superior performance over current state-of-the-art algorithms.
arXiv Detail & Related papers (2021-12-21T02:15:01Z) - A Closer Look at Self-training for Zero-Label Semantic Segmentation [53.4488444382874]
Being able to segment unseen classes not observed during training is an important technical challenge in deep learning.
Prior zero-label semantic segmentation works approach this task by learning visual-semantic embeddings or generative models.
We propose a consistency regularizer to filter out noisy pseudo-labels by taking the intersections of the pseudo-labels generated from different augmentations of the same image.
arXiv Detail & Related papers (2021-04-21T14:34:33Z) - Semantic Segmentation with Generative Models: Semi-Supervised Learning
and Strong Out-of-Domain Generalization [112.68171734288237]
We propose a novel framework for discriminative pixel-level tasks using a generative model of both images and labels.
We learn a generative adversarial network that captures the joint image-label distribution and is trained efficiently using a large set of unlabeled images.
We demonstrate strong in-domain performance compared to several baselines, and are the first to showcase extreme out-of-domain generalization.
arXiv Detail & Related papers (2021-04-12T21:41:25Z) - Unsupervised Person Re-identification via Multi-label Classification [55.65870468861157]
This paper formulates unsupervised person ReID as a multi-label classification task to progressively seek true labels.
Our method starts by assigning each person image with a single-class label, then evolves to multi-label classification by leveraging the updated ReID model for label prediction.
To boost the ReID model training efficiency in multi-label classification, we propose the memory-based multi-label classification loss (MMCL)
arXiv Detail & Related papers (2020-04-20T12:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.