A New Dataset and A Baseline Model for Breast Lesion Detection in
Ultrasound Videos
- URL: http://arxiv.org/abs/2207.00141v1
- Date: Fri, 1 Jul 2022 01:37:50 GMT
- Title: A New Dataset and A Baseline Model for Breast Lesion Detection in
Ultrasound Videos
- Authors: Zhi Lin, Junhao Lin, Lei Zhu, Huazhu Fu, Jing Qin, Liansheng Wang
- Abstract summary: We first collect and annotate an ultrasound video dataset (188 videos) for breast lesion detection.
We propose a clip-level and video-level feature aggregated network (CVA-Net) for addressing breast lesion detection in ultrasound videos.
- Score: 43.42513012531214
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Breast lesion detection in ultrasound is critical for breast cancer
diagnosis. Existing methods mainly rely on individual 2D ultrasound images or
combine unlabeled video and labeled 2D images to train models for breast lesion
detection. In this paper, we first collect and annotate an ultrasound video
dataset (188 videos) for breast lesion detection. Moreover, we propose a
clip-level and video-level feature aggregated network (CVA-Net) for addressing
breast lesion detection in ultrasound videos by aggregating video-level lesion
classification features and clip-level temporal features. The clip-level
temporal features encode local temporal information of ordered video frames and
global temporal information of shuffled video frames. In our CVA-Net, an
inter-video fusion module is devised to fuse local features from original video
frames and global features from shuffled video frames, and an intra-video
fusion module is devised to learn the temporal information among adjacent video
frames. Moreover, we learn video-level features to classify the breast lesions
of the original video as benign or malignant lesions to further enhance the
final breast lesion detection performance in ultrasound videos. Experimental
results on our annotated dataset demonstrate that our CVA-Net clearly
outperforms state-of-the-art methods. The corresponding code and dataset are
publicly available at \url{https://github.com/jhl-Det/CVA-Net}.
Related papers
- Weakly Supervised Video Anomaly Detection and Localization with Spatio-Temporal Prompts [57.01985221057047]
This paper introduces a novel method that learnstemporal prompt embeddings for weakly supervised video anomaly detection and localization (WSVADL) based on pre-trained vision-language models (VLMs)
Our method achieves state-of-theart performance on three public benchmarks for the WSVADL task.
arXiv Detail & Related papers (2024-08-12T03:31:29Z) - A Spatial-Temporal Progressive Fusion Network for Breast Lesion Segmentation in Ultrasound Videos [7.0117363464728815]
The main challenge for ultrasound video-based breast lesion segmentation is how to exploit the lesion cues of both intraframe and inter-frame simultaneously.
We propose a novel Spatial Progressive Fusion Network (STPFNet) for video based breast lesion segmentation problem.
STPFNet achieves better breast lesion detection performance than state-of-the-art methods.
arXiv Detail & Related papers (2024-03-18T11:56:32Z) - Shifting More Attention to Breast Lesion Segmentation in Ultrasound
Videos [43.454994341021276]
We meticulously curated a US video breast lesion segmentation dataset comprising 572 videos and 34,300 annotated frames.
We propose a novel frequency and localization feature aggregation network (FLA-Net) that learns temporal features from the frequency domain.
Our experiments on our annotated dataset and two public video polyp segmentation datasets demonstrate that our proposed FLA-Net achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-10-03T07:50:32Z) - A Spatial-Temporal Deformable Attention based Framework for Breast
Lesion Detection in Videos [107.96514633713034]
We propose a spatial-temporal deformable attention based framework, named STNet.
Our STNet introduces a spatial-temporal deformable attention module to perform local spatial-temporal feature fusion.
Experiments on the public breast lesion ultrasound video dataset show that our STNet obtains a state-of-the-art detection performance.
arXiv Detail & Related papers (2023-09-09T07:00:10Z) - OnUVS: Online Feature Decoupling Framework for High-Fidelity Ultrasound
Video Synthesis [34.07625938756013]
Sonographers must observe corresponding dynamic anatomic structures to gather comprehensive information.
The synthesis of US videos may represent a promising solution to this issue.
We present a novel online feature-decoupling framework called OnUVS for high-fidelity US video synthesis.
arXiv Detail & Related papers (2023-08-16T10:16:50Z) - Boosting Breast Ultrasound Video Classification by the Guidance of
Keyframe Feature Centers [16.527815681294534]
We propose KGA-Net and coherence loss to enhance the performance of ultrasound video classification.
Our method boosts the performance on the public BUSV dataset by a large margin.
arXiv Detail & Related papers (2023-06-12T05:30:09Z) - Spatial-Temporal Frequency Forgery Clue for Video Forgery Detection in
VIS and NIR Scenario [87.72258480670627]
Existing face forgery detection methods based on frequency domain find that the GAN forged images have obvious grid-like visual artifacts in the frequency spectrum compared to the real images.
This paper proposes a Cosine Transform-based Forgery Clue Augmentation Network (FCAN-DCT) to achieve a more comprehensive spatial-temporal feature representation.
arXiv Detail & Related papers (2022-07-05T09:27:53Z) - Global Guidance Network for Breast Lesion Segmentation in Ultrasound
Images [84.03487786163781]
We develop a deep convolutional neural network equipped with a global guidance block (GGB) and breast lesion boundary detection modules.
Our network outperforms other medical image segmentation methods and the recent semantic segmentation methods on breast ultrasound lesion segmentation.
arXiv Detail & Related papers (2021-04-05T13:15:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.