Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network
- URL: http://arxiv.org/abs/2308.02101v1
- Date: Fri, 4 Aug 2023 01:19:32 GMT
- Title: Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network
- Authors: Bryar Shareef, Min Xian, Aleksandar Vakanski, Haotian Wang
- Abstract summary: Capturing global contextual information plays a critical role in breast ultrasound (BUS) image classification.
Vision Transformers have an improved capability of capturing global contextual information but may distort the local image patterns due to the tokenization operations.
In this study, we proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN, designed to perform BUS tumor classification and segmentation.
- Score: 63.845552349914186
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Capturing global contextual information plays a critical role in breast
ultrasound (BUS) image classification. Although convolutional neural networks
(CNNs) have demonstrated reliable performance in tumor classification, they
have inherent limitations for modeling global and long-range dependencies due
to the localized nature of convolution operations. Vision Transformers have an
improved capability of capturing global contextual information but may distort
the local image patterns due to the tokenization operations. In this study, we
proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN,
designed to perform BUS tumor classification and segmentation using a hybrid
architecture composed of CNNs and Swin Transformer components. The proposed
approach was compared to nine BUS classification methods and evaluated using
seven quantitative metrics on a dataset of 3,320 BUS images. The results
indicate that Hybrid-MT-ESTAN achieved the highest accuracy, sensitivity, and
F1 score of 82.7%, 86.4%, and 86.0%, respectively.
Related papers
- Attention-Enhanced Hybrid Feature Aggregation Network for 3D Brain Tumor Segmentation [0.9897828700959131]
Glioblastoma is a highly aggressive and malignant brain tumor type that requires early diagnosis and prompt intervention.
To address this challenge, Artificial Intelligence (AI)-driven approaches in healthcare have generated interest in efficiently diagnosing and evaluating brain tumors.
In our approach, we utilize a multi-scale, attention-guided and hybrid U-Net-shaped model -- GLIMS -- to perform 3D brain tumor segmentation.
arXiv Detail & Related papers (2024-03-15T00:52:17Z) - SDR-Former: A Siamese Dual-Resolution Transformer for Liver Lesion
Classification Using 3D Multi-Phase Imaging [59.78761085714715]
This study proposes a novel Siamese Dual-Resolution Transformer (SDR-Former) framework for liver lesion classification.
The proposed framework has been validated through comprehensive experiments on two clinical datasets.
To support the scientific community, we are releasing our extensive multi-phase MR dataset for liver lesion analysis to the public.
arXiv Detail & Related papers (2024-02-27T06:32:56Z) - CAFCT: Contextual and Attentional Feature Fusions of Convolutional
Neural Networks and Transformer for Liver Tumor Segmentation [4.255240258747643]
We propose a Contextual and Attentional feature Fusions enhanced Convolutional Network (CNN) and Transformer hybrid network (CAFCT) model for liver tumor segmentation.
Experimental results show that the proposed CAFCT achieves semantic Intersection of 90.38% and Dice score of 86.78%, respectively.
arXiv Detail & Related papers (2024-01-30T10:42:11Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - How GNNs Facilitate CNNs in Mining Geometric Information from
Large-Scale Medical Images [2.2699159408903484]
We propose a fusion framework for enhancing the global image-level representation captured by convolutional neural networks (CNNs)
We evaluate our fusion strategies on histology datasets curated from large patient cohorts of colorectal and gastric cancers.
arXiv Detail & Related papers (2022-06-15T15:27:48Z) - EMT-NET: Efficient multitask network for computer-aided diagnosis of
breast cancer [58.720142291102135]
We propose an efficient and light-weighted learning architecture to classify and segment breast tumors simultaneously.
We incorporate a segmentation task into a tumor classification network, which makes the backbone network learn representations focused on tumor regions.
The accuracy, sensitivity, and specificity of tumor classification is 88.6%, 94.1%, and 85.3%, respectively.
arXiv Detail & Related papers (2022-01-13T05:24:40Z) - BiTr-Unet: a CNN-Transformer Combined Network for MRI Brain Tumor
Segmentation [2.741266294612776]
We present a CNN-Transformer combined model called BiTr-Unet for brain tumor segmentation on multi-modal MRI scans.
The proposed BiTr-Unet achieves good performance on the BraTS 2021 validation dataset with mean Dice score 0.9076, 0.8392 and 0.8231, and mean Hausdorff distance 4.5322, 13.4592 and 14.9963 for the whole tumor, tumor core, and enhancing tumor, respectively.
arXiv Detail & Related papers (2021-09-25T04:18:34Z) - Medulloblastoma Tumor Classification using Deep Transfer Learning with
Multi-Scale EfficientNets [63.62764375279861]
We propose an end-to-end MB tumor classification and explore transfer learning with various input sizes and matching network dimensions.
Using a data set with 161 cases, we demonstrate that pre-trained EfficientNets with larger input resolutions lead to significant performance improvements.
arXiv Detail & Related papers (2021-09-10T13:07:11Z) - TransMIL: Transformer based Correlated Multiple Instance Learning for
Whole Slide Image Classication [38.58585442160062]
Multiple instance learning (MIL) is a powerful tool to solve the weakly supervised classification in whole slide image (WSI) based pathology diagnosis.
We proposed a new framework, called correlated MIL, and provided a proof for convergence.
We conducted various experiments for three different computational pathology problems and achieved better performance and faster convergence compared with state-of-the-art methods.
arXiv Detail & Related papers (2021-06-02T02:57:54Z) - Comparisons among different stochastic selection of activation layers
for convolutional neural networks for healthcare [77.99636165307996]
We classify biomedical images using ensembles of neural networks.
We select our activations among the following ones: ReLU, leaky ReLU, Parametric ReLU, ELU, Adaptive Piecewice Linear Unit, S-Shaped ReLU, Swish, Mish, Mexican Linear Unit, Parametric Deformable Linear Unit, Soft Root Sign.
arXiv Detail & Related papers (2020-11-24T01:53:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.