Is Two-shot All You Need? A Label-efficient Approach for Video
Segmentation in Breast Ultrasound
- URL: http://arxiv.org/abs/2402.04921v2
- Date: Mon, 4 Mar 2024 02:41:21 GMT
- Title: Is Two-shot All You Need? A Label-efficient Approach for Video
Segmentation in Breast Ultrasound
- Authors: Jiajun Zeng, Dong Ni, Ruobing Huang
- Abstract summary: We propose a novel two-shot training paradigm for BUS video segmentation.
It not only is able to capture free-range space-time consistency but also utilizes a source-dependent augmentation scheme.
Results showed that it gained comparable performance to the fully annotated ones given only 1.9% training labels.
- Score: 4.113689581316844
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Breast lesion segmentation from breast ultrasound (BUS) videos could assist
in early diagnosis and treatment. Existing video object segmentation (VOS)
methods usually require dense annotation, which is often inaccessible for
medical datasets. Furthermore, they suffer from accumulative errors and a lack
of explicit space-time awareness. In this work, we propose a novel two-shot
training paradigm for BUS video segmentation. It not only is able to capture
free-range space-time consistency but also utilizes a source-dependent
augmentation scheme. This label-efficient learning framework is validated on a
challenging in-house BUS video dataset. Results showed that it gained
comparable performance to the fully annotated ones given only 1.9% training
labels.
Related papers
- A Spatial-Temporal Progressive Fusion Network for Breast Lesion Segmentation in Ultrasound Videos [7.0117363464728815]
The main challenge for ultrasound video-based breast lesion segmentation is how to exploit the lesion cues of both intraframe and inter-frame simultaneously.
We propose a novel Spatial Progressive Fusion Network (STPFNet) for video based breast lesion segmentation problem.
STPFNet achieves better breast lesion detection performance than state-of-the-art methods.
arXiv Detail & Related papers (2024-03-18T11:56:32Z) - Semi- and Weakly-Supervised Learning for Mammogram Mass Segmentation with Limited Annotations [49.33388736227072]
We propose a semi- and weakly-supervised learning framework for mass segmentation.
We use limited strongly-labeled samples and sufficient weakly-labeled samples to achieve satisfactory performance.
Experiments on CBIS-DDSM and INbreast datasets demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2024-03-14T12:05:25Z) - Vivim: a Video Vision Mamba for Medical Video Segmentation [52.11785024350253]
This paper presents a Video Vision Mamba-based framework, dubbed as Vivim, for medical video segmentation tasks.
Our Vivim can effectively compress the long-term representation into sequences at varying scales.
Experiments on thyroid segmentation, breast lesion segmentation in ultrasound videos, and polyp segmentation in colonoscopy videos demonstrate the effectiveness and efficiency of our Vivim.
arXiv Detail & Related papers (2024-01-25T13:27:03Z) - Shifting More Attention to Breast Lesion Segmentation in Ultrasound
Videos [43.454994341021276]
We meticulously curated a US video breast lesion segmentation dataset comprising 572 videos and 34,300 annotated frames.
We propose a novel frequency and localization feature aggregation network (FLA-Net) that learns temporal features from the frequency domain.
Our experiments on our annotated dataset and two public video polyp segmentation datasets demonstrate that our proposed FLA-Net achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-10-03T07:50:32Z) - A Simple Framework for Open-Vocabulary Segmentation and Detection [85.21641508535679]
We present OpenSeeD, a simple Open-vocabulary and Detection framework that jointly learns from different segmentation and detection datasets.
We first introduce a pre-trained text encoder to encode all the visual concepts in two tasks and learn a common semantic space for them.
After pre-training, our model exhibits competitive or stronger zero-shot transferability for both segmentation and detection.
arXiv Detail & Related papers (2023-03-14T17:58:34Z) - Self-Supervised Correction Learning for Semi-Supervised Biomedical Image
Segmentation [84.58210297703714]
We propose a self-supervised correction learning paradigm for semi-supervised biomedical image segmentation.
We design a dual-task network, including a shared encoder and two independent decoders for segmentation and lesion region inpainting.
Experiments on three medical image segmentation datasets for different tasks demonstrate the outstanding performance of our method.
arXiv Detail & Related papers (2023-01-12T08:19:46Z) - Pseudo-label Guided Cross-video Pixel Contrast for Robotic Surgical
Scene Segmentation with Limited Annotations [72.15956198507281]
We propose PGV-CL, a novel pseudo-label guided cross-video contrast learning method to boost scene segmentation.
We extensively evaluate our method on a public robotic surgery dataset EndoVis18 and a public cataract dataset CaDIS.
arXiv Detail & Related papers (2022-07-20T05:42:19Z) - A New Dataset and A Baseline Model for Breast Lesion Detection in
Ultrasound Videos [43.42513012531214]
We first collect and annotate an ultrasound video dataset (188 videos) for breast lesion detection.
We propose a clip-level and video-level feature aggregated network (CVA-Net) for addressing breast lesion detection in ultrasound videos.
arXiv Detail & Related papers (2022-07-01T01:37:50Z) - Weakly Supervised Contrastive Learning for Better Severity Scoring of
Lung Ultrasound [0.044364554283083675]
Several AI-based patient severity scoring models have been proposed that rely on scoring the appearance of the ultrasound scans.
We address the challenge of labeling every ultrasound frame in the video clips.
Our contrastive learning method treats the video clip severity labels as noisy weak severity labels for individual frames.
We show that it performs better than the conventional cross-entropy loss based training.
arXiv Detail & Related papers (2022-01-18T23:45:18Z) - ATSO: Asynchronous Teacher-Student Optimization for Semi-Supervised
Medical Image Segmentation [99.90263375737362]
We propose ATSO, an asynchronous version of teacher-student optimization.
ATSO partitions the unlabeled data into two subsets and alternately uses one subset to fine-tune the model and updates the label on the other subset.
We evaluate ATSO on two popular medical image segmentation datasets and show its superior performance in various semi-supervised settings.
arXiv Detail & Related papers (2020-06-24T04:05:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.