Breast Lesion Diagnosis Using Static Images and Dynamic Video
- URL: http://arxiv.org/abs/2308.09980v1
- Date: Sat, 19 Aug 2023 11:09:58 GMT
- Title: Breast Lesion Diagnosis Using Static Images and Dynamic Video
- Authors: Yunwen Huang, Hongyu Hu, Ying Zhu, Yi Xu
- Abstract summary: We propose a multi-modality breast tumor diagnosis model to imitate the diagnosing process of radiologists.
Our work is validated on a breast ultrasound dataset composed of 897 sets of ultrasound images and videos.
- Score: 12.71602984461284
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning based Computer Aided Diagnosis (CAD) systems have been
developed to treat breast ultrasound. Most of them focus on a single ultrasound
imaging modality, either using representative static images or the dynamic
video of a real-time scan. In fact, these two image modalities are
complementary for lesion diagnosis. Dynamic videos provide detailed
three-dimensional information about the lesion, while static images capture the
typical sections of the lesion. In this work, we propose a multi-modality
breast tumor diagnosis model to imitate the diagnosing process of radiologists,
which learns the features of both static images and dynamic video and explores
the potential relationship between the two modalities. Considering that static
images are carefully selected by professional radiologists, we propose to
aggregate dynamic video features under the guidance of domain knowledge from
static images before fusing multi-modality features. Our work is validated on a
breast ultrasound dataset composed of 897 sets of ultrasound images and videos.
Experimental results show that our model boosts the performance of
Benign/Malignant classification, achieving 90.0% in AUC and 81.7% in accuracy.
Related papers
- Breast tumor classification based on self-supervised contrastive learning from ultrasound videos [7.825379326219145]
We adopted a triplet network and a self-supervised contrastive learning technique to learn representations from unlabeled breast ultrasound video clips.
Our model achieved an area under the receiver operating characteristic curve (AUC) of 0.952, which is significantly higher than the others.
The proposed framework greatly reduces the demand for labeled data and holds potential for use in automatic breast ultrasound image diagnosis.
arXiv Detail & Related papers (2024-08-20T07:16:01Z) - Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - OnUVS: Online Feature Decoupling Framework for High-Fidelity Ultrasound
Video Synthesis [34.07625938756013]
Sonographers must observe corresponding dynamic anatomic structures to gather comprehensive information.
The synthesis of US videos may represent a promising solution to this issue.
We present a novel online feature-decoupling framework called OnUVS for high-fidelity US video synthesis.
arXiv Detail & Related papers (2023-08-16T10:16:50Z) - On Sensitivity and Robustness of Normalization Schemes to Input
Distribution Shifts in Automatic MR Image Diagnosis [58.634791552376235]
Deep Learning (DL) models have achieved state-of-the-art performance in diagnosing multiple diseases using reconstructed images as input.
DL models are sensitive to varying artifacts as it leads to changes in the input data distribution between the training and testing phases.
We propose to use other normalization techniques, such as Group Normalization and Layer Normalization, to inject robustness into model performance against varying image artifacts.
arXiv Detail & Related papers (2023-06-23T03:09:03Z) - XrayGPT: Chest Radiographs Summarization using Medical Vision-Language
Models [60.437091462613544]
We introduce XrayGPT, a novel conversational medical vision-language model.
It can analyze and answer open-ended questions about chest radiographs.
We generate 217k interactive and high-quality summaries from free-text radiology reports.
arXiv Detail & Related papers (2023-06-13T17:59:59Z) - Using Spatio-Temporal Dual-Stream Network with Self-Supervised Learning
for Lung Tumor Classification on Radial Probe Endobronchial Ultrasound Video [0.0]
During the biopsy process of lung cancer, physicians use real-time ultrasound images to find suitable lesion locations for sampling.
Previous studies have employed 2D convolutional neural networks to effectively differentiate between benign and malignant lung lesions.
This study designs an automatic diagnosis system based on a 3D neural network.
arXiv Detail & Related papers (2023-05-04T10:39:37Z) - Video4MRI: An Empirical Study on Brain Magnetic Resonance Image
Analytics with CNN-based Video Classification Frameworks [60.42012344842292]
3D CNN-based models dominate the field of magnetic resonance image (MRI) analytics.
In this paper, four datasets of Alzheimer's and Parkinson's disease recognition are utilized in experiments.
In terms of efficiency, the video framework performs better than 3D-CNN models by 5% - 11% with 50% - 66% less trainable parameters.
arXiv Detail & Related papers (2023-02-24T15:26:31Z) - Self-supervised Learning from 100 Million Medical Images [13.958840691105992]
We propose a method for self-supervised learning of rich image features based on contrastive learning and online feature clustering.
We leverage large training datasets of over 100,000,000 medical images of various modalities, including radiography, computed tomography (CT), magnetic resonance (MR) imaging and ultrasonography.
We highlight a number of advantages of this strategy on challenging image assessment problems in radiography, CT and MR.
arXiv Detail & Related papers (2022-01-04T18:27:04Z) - Voice-assisted Image Labelling for Endoscopic Ultrasound Classification
using Neural Networks [48.732863591145964]
We propose a multi-modal convolutional neural network architecture that labels endoscopic ultrasound (EUS) images from raw verbal comments provided by a clinician during the procedure.
Our results show a prediction accuracy of 76% at image level on a dataset with 5 different labels.
arXiv Detail & Related papers (2021-10-12T21:22:24Z) - Deep Learning for Ultrasound Beamforming [120.12255978513912]
Beamforming, the process of mapping received ultrasound echoes to the spatial image domain, lies at the heart of the ultrasound image formation chain.
Modern ultrasound imaging leans heavily on innovations in powerful digital receive channel processing.
Deep learning methods can play a compelling role in the digital beamforming pipeline.
arXiv Detail & Related papers (2021-09-23T15:15:21Z) - Ensemble Transfer Learning of Elastography and B-mode Breast Ultrasound
Images [3.3615086420912745]
We present an ensemble transfer learning model to classify benign and malignant breast tumors.
This model combines semantic features from AlexNet & ResNet models to classify benign from malignant tumors.
Experimental results show that our ensemble model achieves a sensitivity of 88.89% and specificity of 91.10%.
arXiv Detail & Related papers (2021-02-17T04:23:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.