Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network
- URL: http://arxiv.org/abs/2308.02101v1
- Date: Fri, 4 Aug 2023 01:19:32 GMT
- Title: Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network
- Authors: Bryar Shareef, Min Xian, Aleksandar Vakanski, Haotian Wang
- Abstract summary: Capturing global contextual information plays a critical role in breast ultrasound (BUS) image classification.
Vision Transformers have an improved capability of capturing global contextual information but may distort the local image patterns due to the tokenization operations.
In this study, we proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN, designed to perform BUS tumor classification and segmentation.
- Score: 63.845552349914186
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Capturing global contextual information plays a critical role in breast
ultrasound (BUS) image classification. Although convolutional neural networks
(CNNs) have demonstrated reliable performance in tumor classification, they
have inherent limitations for modeling global and long-range dependencies due
to the localized nature of convolution operations. Vision Transformers have an
improved capability of capturing global contextual information but may distort
the local image patterns due to the tokenization operations. In this study, we
proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN,
designed to perform BUS tumor classification and segmentation using a hybrid
architecture composed of CNNs and Swin Transformer components. The proposed
approach was compared to nine BUS classification methods and evaluated using
seven quantitative metrics on a dataset of 3,320 BUS images. The results
indicate that Hybrid-MT-ESTAN achieved the highest accuracy, sensitivity, and
F1 score of 82.7%, 86.4%, and 86.0%, respectively.
Related papers
- Advanced Hybrid Deep Learning Model for Enhanced Classification of Osteosarcoma Histopathology Images [0.0]
This study focuses on osteosarcoma (OS), the most common bone cancer in children and adolescents, which affects the long bones of the arms and legs.
We propose a novel hybrid model that combines convolutional neural networks (CNN) and vision transformers (ViT) to improve diagnostic accuracy for OS.
The model achieved an accuracy of 99.08%, precision of 99.10%, recall of 99.28%, and an F1-score of 99.23%.
arXiv Detail & Related papers (2024-10-29T13:54:08Z) - Prototype Learning Guided Hybrid Network for Breast Tumor Segmentation in DCE-MRI [58.809276442508256]
We propose a hybrid network via the combination of convolution neural network (CNN) and transformer layers.
The experimental results on private and public DCE-MRI datasets demonstrate that the proposed hybrid network superior performance than the state-of-the-art methods.
arXiv Detail & Related papers (2024-08-11T15:46:00Z) - SDR-Former: A Siamese Dual-Resolution Transformer for Liver Lesion
Classification Using 3D Multi-Phase Imaging [59.78761085714715]
This study proposes a novel Siamese Dual-Resolution Transformer (SDR-Former) framework for liver lesion classification.
The proposed framework has been validated through comprehensive experiments on two clinical datasets.
To support the scientific community, we are releasing our extensive multi-phase MR dataset for liver lesion analysis to the public.
arXiv Detail & Related papers (2024-02-27T06:32:56Z) - CAFCT-Net: A CNN-Transformer Hybrid Network with Contextual and Attentional Feature Fusion for Liver Tumor Segmentation [3.8952128960495638]
We propose a Contextual and Attentional feature Fusions enhanced Convolutional Network (CNN) and Transformer hybrid network (CAFCT-Net) for liver tumor segmentation.
Experimental results show that the proposed model achieves a mean Intersection of 76.54% and Dice coefficient of 84.29%, respectively.
arXiv Detail & Related papers (2024-01-30T10:42:11Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - EMT-NET: Efficient multitask network for computer-aided diagnosis of
breast cancer [58.720142291102135]
We propose an efficient and light-weighted learning architecture to classify and segment breast tumors simultaneously.
We incorporate a segmentation task into a tumor classification network, which makes the backbone network learn representations focused on tumor regions.
The accuracy, sensitivity, and specificity of tumor classification is 88.6%, 94.1%, and 85.3%, respectively.
arXiv Detail & Related papers (2022-01-13T05:24:40Z) - BiTr-Unet: a CNN-Transformer Combined Network for MRI Brain Tumor
Segmentation [2.741266294612776]
We present a CNN-Transformer combined model called BiTr-Unet for brain tumor segmentation on multi-modal MRI scans.
The proposed BiTr-Unet achieves good performance on the BraTS 2021 validation dataset with mean Dice score 0.9076, 0.8392 and 0.8231, and mean Hausdorff distance 4.5322, 13.4592 and 14.9963 for the whole tumor, tumor core, and enhancing tumor, respectively.
arXiv Detail & Related papers (2021-09-25T04:18:34Z) - Medulloblastoma Tumor Classification using Deep Transfer Learning with
Multi-Scale EfficientNets [63.62764375279861]
We propose an end-to-end MB tumor classification and explore transfer learning with various input sizes and matching network dimensions.
Using a data set with 161 cases, we demonstrate that pre-trained EfficientNets with larger input resolutions lead to significant performance improvements.
arXiv Detail & Related papers (2021-09-10T13:07:11Z) - TransMIL: Transformer based Correlated Multiple Instance Learning for
Whole Slide Image Classication [38.58585442160062]
Multiple instance learning (MIL) is a powerful tool to solve the weakly supervised classification in whole slide image (WSI) based pathology diagnosis.
We proposed a new framework, called correlated MIL, and provided a proof for convergence.
We conducted various experiments for three different computational pathology problems and achieved better performance and faster convergence compared with state-of-the-art methods.
arXiv Detail & Related papers (2021-06-02T02:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.