Ensemble Learning with Residual Transformer for Brain Tumor Segmentation
- URL: http://arxiv.org/abs/2308.00128v1
- Date: Mon, 31 Jul 2023 19:47:33 GMT
- Title: Ensemble Learning with Residual Transformer for Brain Tumor Segmentation
- Authors: Lanhong Yao, Zheyuan Zhang, Ulas Bagci
- Abstract summary: This paper proposes a novel network architecture that integrates Transformers into a self-adaptive U-Net.
On the BraTS 2021 dataset (3D), our model achieves 87.6% mean Dice score and outperforms the state-of-the-art methods.
- Score: 2.0654955576087084
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Brain tumor segmentation is an active research area due to the difficulty in
delineating highly complex shaped and textured tumors as well as the failure of
the commonly used U-Net architectures. The combination of different neural
architectures is among the mainstream research recently, particularly the
combination of U-Net with Transformers because of their innate attention
mechanism and pixel-wise labeling. Different from previous efforts, this paper
proposes a novel network architecture that integrates Transformers into a
self-adaptive U-Net to draw out 3D volumetric contexts with reasonable
computational costs. We further add a residual connection to prevent
degradation in information flow and explore ensemble methods, as the evaluated
models have edges on different cases and sub-regions. On the BraTS 2021 dataset
(3D), our model achieves 87.6% mean Dice score and outperforms the
state-of-the-art methods, demonstrating the potential for combining multiple
architectures to optimize brain tumor segmentation.
Related papers
- MBDRes-U-Net: Multi-Scale Lightweight Brain Tumor Segmentation Network [0.0]
This study proposes the MBDRes-U-Net model using the three-dimensional (3D) U-Net framework, which integrates multibranch residual blocks and fused attention into the model.
The computational burden of the model is reduced by the branch strategy, which effectively uses the rich local features in multimodal images.
arXiv Detail & Related papers (2024-11-04T09:03:43Z) - CAFCT-Net: A CNN-Transformer Hybrid Network with Contextual and Attentional Feature Fusion for Liver Tumor Segmentation [3.8952128960495638]
We propose a Contextual and Attentional feature Fusions enhanced Convolutional Network (CNN) and Transformer hybrid network (CAFCT-Net) for liver tumor segmentation.
Experimental results show that the proposed model achieves a mean Intersection of 76.54% and Dice coefficient of 84.29%, respectively.
arXiv Detail & Related papers (2024-01-30T10:42:11Z) - Leveraging Frequency Domain Learning in 3D Vessel Segmentation [50.54833091336862]
In this study, we leverage Fourier domain learning as a substitute for multi-scale convolutional kernels in 3D hierarchical segmentation models.
We show that our novel network achieves remarkable dice performance (84.37% on ASACA500 and 80.32% on ImageCAS) in tubular vessel segmentation tasks.
arXiv Detail & Related papers (2024-01-11T19:07:58Z) - Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network [63.845552349914186]
Capturing global contextual information plays a critical role in breast ultrasound (BUS) image classification.
Vision Transformers have an improved capability of capturing global contextual information but may distort the local image patterns due to the tokenization operations.
In this study, we proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN, designed to perform BUS tumor classification and segmentation.
arXiv Detail & Related papers (2023-08-04T01:19:32Z) - 3D Brainformer: 3D Fusion Transformer for Brain Tumor Segmentation [6.127298607534532]
Deep learning has recently emerged to improve brain tumor segmentation.
Transformers have been leveraged to address the limitations of convolutional networks.
We propose a 3D Transformer-based segmentation approach.
arXiv Detail & Related papers (2023-04-28T02:11:29Z) - View-Disentangled Transformer for Brain Lesion Detection [50.4918615815066]
We propose a novel view-disentangled transformer to enhance the extraction of MRI features for more accurate tumour detection.
First, the proposed transformer harvests long-range correlation among different positions in a 3D brain scan.
Second, the transformer models a stack of slice features as multiple 2D views and enhance these features view-by-view.
Third, we deploy the proposed transformer module in a transformer backbone, which can effectively detect the 2D regions surrounding brain lesions.
arXiv Detail & Related papers (2022-09-20T11:58:23Z) - MISSU: 3D Medical Image Segmentation via Self-distilling TransUNet [55.16833099336073]
We propose to self-distill a Transformer-based UNet for medical image segmentation.
It simultaneously learns global semantic information and local spatial-detailed features.
Our MISSU achieves the best performance over previous state-of-the-art methods.
arXiv Detail & Related papers (2022-06-02T07:38:53Z) - Automatic size and pose homogenization with spatial transformer network
to improve and accelerate pediatric segmentation [51.916106055115755]
We propose a new CNN architecture that is pose and scale invariant thanks to the use of Spatial Transformer Network (STN)
Our architecture is composed of three sequential modules that are estimated together during training.
We test the proposed method in kidney and renal tumor segmentation on abdominal pediatric CT scanners.
arXiv Detail & Related papers (2021-07-06T14:50:03Z) - CoTr: Efficiently Bridging CNN and Transformer for 3D Medical Image
Segmentation [95.51455777713092]
Convolutional neural networks (CNNs) have been the de facto standard for nowadays 3D medical image segmentation.
We propose a novel framework that efficiently bridges a bf Convolutional neural network and a bf Transformer bf (CoTr) for accurate 3D medical image segmentation.
arXiv Detail & Related papers (2021-03-04T13:34:22Z) - HI-Net: Hyperdense Inception 3D UNet for Brain Tumor Segmentation [17.756591105686]
This paper proposes hyperdense inception 3D UNet (HI-Net), which captures multi-scale information by stacking factorization of 3D weighted convolutional layers in the residual inception block.
Preliminary results on the BRATS 2020 testing set show that achieved by our proposed approach, the dice (DSC) scores of ET, WT, and TC are 0.79457, 0.87494, and 0.83712, respectively.
arXiv Detail & Related papers (2020-12-12T09:09:04Z) - Brain tumour segmentation using cascaded 3D densely-connected U-net [10.667165962654996]
We propose a deep-learning based method to segment a brain tumour into its subregions.
The proposed architecture is a 3D convolutional neural network based on a variant of the U-Net architecture.
Experimental results on the BraTS20 validation dataset demonstrate that the proposed model achieved average Dice Scores of 0.90, 0.82, and 0.78 for whole tumour, tumour core and enhancing tumour respectively.
arXiv Detail & Related papers (2020-09-16T09:14:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.