MUVF-YOLOX: A Multi-modal Ultrasound Video Fusion Network for Renal
Tumor Diagnosis
- URL: http://arxiv.org/abs/2307.07807v1
- Date: Sat, 15 Jul 2023 14:15:42 GMT
- Title: MUVF-YOLOX: A Multi-modal Ultrasound Video Fusion Network for Renal
Tumor Diagnosis
- Authors: Junyu Li, Han Huang, Dong Ni, Wufeng Xue, Dongmei Zhu, Jun Cheng
- Abstract summary: We propose a novel multi-modal ultrasound video fusion network that can effectively perform multi-modal feature fusion and video classification for renal tumor diagnosis.
Experimental results on a multicenter dataset show that the proposed framework outperforms the single-modal models and the competing methods.
- Score: 10.452919030855796
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Early diagnosis of renal cancer can greatly improve the survival rate of
patients. Contrast-enhanced ultrasound (CEUS) is a cost-effective and
non-invasive imaging technique and has become more and more frequently used for
renal tumor diagnosis. However, the classification of benign and malignant
renal tumors can still be very challenging due to the highly heterogeneous
appearance of cancer and imaging artifacts. Our aim is to detect and classify
renal tumors by integrating B-mode and CEUS-mode ultrasound videos. To this
end, we propose a novel multi-modal ultrasound video fusion network that can
effectively perform multi-modal feature fusion and video classification for
renal tumor diagnosis. The attention-based multi-modal fusion module uses
cross-attention and self-attention to extract modality-invariant features and
modality-specific features in parallel. In addition, we design an object-level
temporal aggregation (OTA) module that can automatically filter low-quality
features and efficiently integrate temporal information from multiple frames to
improve the accuracy of tumor diagnosis. Experimental results on a multicenter
dataset show that the proposed framework outperforms the single-modal models
and the competing methods. Furthermore, our OTA module achieves higher
classification accuracy than the frame-level predictions. Our code is available
at \url{https://github.com/JeunyuLi/MUAF}.
Related papers
- Unifying Subsampling Pattern Variations for Compressed Sensing MRI with Neural Operators [72.79532467687427]
Compressed Sensing MRI reconstructs images of the body's internal anatomy from undersampled and compressed measurements.
Deep neural networks have shown great potential for reconstructing high-quality images from highly undersampled measurements.
We propose a unified model that is robust to different subsampling patterns and image resolutions in CS-MRI.
arXiv Detail & Related papers (2024-10-05T20:03:57Z) - Towards a Benchmark for Colorectal Cancer Segmentation in Endorectal Ultrasound Videos: Dataset and Model Development [59.74920439478643]
In this paper, we collect and annotated the first benchmark dataset that covers diverse ERUS scenarios.
Our ERUS-10K dataset comprises 77 videos and 10,000 high-resolution annotated frames.
We introduce a benchmark model for colorectal cancer segmentation, named the Adaptive Sparse-context TRansformer (ASTR)
arXiv Detail & Related papers (2024-08-19T15:04:42Z) - MMFusion: Multi-modality Diffusion Model for Lymph Node Metastasis Diagnosis in Esophageal Cancer [13.74067035373274]
We introduce a multi-modal heterogeneous graph-based conditional feature-guided diffusion model for lymph node metastasis diagnosis based on CT images.
We propose a masked relational representation learning strategy, aiming to uncover the latent prognostic correlations and priorities of primary tumor and lymph node image representations.
arXiv Detail & Related papers (2024-05-15T17:52:00Z) - Cross-modality Guidance-aided Multi-modal Learning with Dual Attention
for MRI Brain Tumor Grading [47.50733518140625]
Brain tumor represents one of the most fatal cancers around the world, and is very common in children and the elderly.
We propose a novel cross-modality guidance-aided multi-modal learning with dual attention for addressing the task of MRI brain tumor grading.
arXiv Detail & Related papers (2024-01-17T07:54:49Z) - Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification [74.84221280249876]
An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
arXiv Detail & Related papers (2022-05-08T15:29:54Z) - RCA-IUnet: A residual cross-spatial attention guided inception U-Net
model for tumor segmentation in breast ultrasound imaging [0.6091702876917281]
The article introduces an efficient residual cross-spatial attention guided inception U-Net (RCA-IUnet) model with minimal training parameters for tumor segmentation.
The RCA-IUnet model follows U-Net topology with residual inception depth-wise separable convolution and hybrid pooling layers.
Cross-spatial attention filters are added to suppress the irrelevant features and focus on the target structure.
arXiv Detail & Related papers (2021-08-05T10:35:06Z) - Modality-aware Mutual Learning for Multi-modal Medical Image
Segmentation [12.308579499188921]
Liver cancer is one of the most common cancers worldwide.
In this paper, we focus on improving automated liver tumor segmentation by integrating multi-modal CT images.
We propose a novel mutual learning (ML) strategy for effective and robust liver tumor segmentation.
arXiv Detail & Related papers (2021-07-21T02:24:31Z) - Learned super resolution ultrasound for improved breast lesion
characterization [52.77024349608834]
Super resolution ultrasound localization microscopy enables imaging of the microvasculature at the capillary level.
In this work we use a deep neural network architecture that makes effective use of signal structure to address these challenges.
By leveraging our trained network, the microvasculature structure is recovered in a short time, without prior PSF knowledge, and without requiring separability of the UCAs.
arXiv Detail & Related papers (2021-07-12T09:04:20Z) - Auto-weighting for Breast Cancer Classification in Multimodal Ultrasound [0.0]
We propose an automatic way to combine the four types of ultrasonography to discriminate between benign and malignant breast nodules.
A novel multimodal network is proposed, along with promising learnability and simplicity to improve classification accuracy.
Results showed that the model scored a high classification accuracy of 95.4%, which indicates the efficiency of the proposed method.
arXiv Detail & Related papers (2020-08-08T03:42:00Z) - Spectral-Spatial Recurrent-Convolutional Networks for In-Vivo
Hyperspectral Tumor Type Classification [49.32653090178743]
We demonstrate the feasibility of in-vivo tumor type classification using hyperspectral imaging and deep learning.
Our best model achieves an AUC of 76.3%, significantly outperforming previous conventional and deep learning methods.
arXiv Detail & Related papers (2020-07-02T12:00:53Z) - A Novel and Efficient Tumor Detection Framework for Pancreatic Cancer
via CT Images [21.627818410241552]
A novel and efficient pancreatic tumor detection framework is proposed in this paper.
The contribution of the proposed method mainly consists of three components: Augmented Feature Pyramid networks, Self-adaptive Feature Fusion and a Dependencies Computation Module.
Experimental results achieve competitive performance in detection with the AUC of 0.9455, which outperforms other state-of-the-art methods to our best of knowledge.
arXiv Detail & Related papers (2020-02-11T15:48:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.