Residual Vision Transformer (ResViT) Based Self-Supervised Learning Model for Brain Tumor Classification
- URL: http://arxiv.org/abs/2411.12874v1
- Date: Tue, 19 Nov 2024 21:42:57 GMT
- Title: Residual Vision Transformer (ResViT) Based Self-Supervised Learning Model for Brain Tumor Classification
- Authors: Meryem Altin Karagoz, O. Ufuk Nalbantoglu, Geoffrey C. Fox,
- Abstract summary: Self-supervised learning models provide data-efficient and remarkable solutions to limited dataset problems.
This paper introduces a generative SSL model for brain tumor classification in two stages.
The proposed model attains the highest accuracy, achieving 90.56% on the BraTs dataset with T1 sequence, 98.53% on the Figshare, and 98.47% on the Kaggle brain tumor datasets.
- Score: 0.08192907805418585
- License:
- Abstract: Deep learning has proven very promising for interpreting MRI in brain tumor diagnosis. However, deep learning models suffer from a scarcity of brain MRI datasets for effective training. Self-supervised learning (SSL) models provide data-efficient and remarkable solutions to limited dataset problems. Therefore, this paper introduces a generative SSL model for brain tumor classification in two stages. The first stage is designed to pre-train a Residual Vision Transformer (ResViT) model for MRI synthesis as a pretext task. The second stage includes fine-tuning a ResViT-based classifier model as a downstream task. Accordingly, we aim to leverage local features via CNN and global features via ViT, employing a hybrid CNN-transformer architecture for ResViT in pretext and downstream tasks. Moreover, synthetic MRI images are utilized to balance the training set. The proposed model performs on public BraTs 2023, Figshare, and Kaggle datasets. Furthermore, we compare the proposed model with various deep learning models, including A-UNet, ResNet-9, pix2pix, pGAN for MRI synthesis, and ConvNeXtTiny, ResNet101, DenseNet12, Residual CNN, ViT for classification. According to the results, the proposed model pretraining on the MRI dataset is superior compared to the pretraining on the ImageNet dataset. Overall, the proposed model attains the highest accuracy, achieving 90.56% on the BraTs dataset with T1 sequence, 98.53% on the Figshare, and 98.47% on the Kaggle brain tumor datasets. As a result, the proposed model demonstrates a robust, effective, and successful approach to handling insufficient dataset challenges in MRI analysis by incorporating SSL, fine-tuning, data augmentation, and combining CNN and ViT.
Related papers
- Brain Tumor Classification on MRI in Light of Molecular Markers [61.77272414423481]
Co-deletion of the 1p/19q gene is associated with clinical outcomes in low-grade gliomas.
This study aims to utilize a specially MRI-based convolutional neural network for brain cancer detection.
arXiv Detail & Related papers (2024-09-29T07:04:26Z) - Predicting recovery following stroke: deep learning, multimodal data and
feature selection using explainable AI [3.797471910783104]
Major challenges include the very high dimensionality of neuroimaging data and the relatively small size of the datasets available for learning.
We introduce a novel approach of training a convolutional neural network (CNN) on images that combine regions-of-interest extracted from MRIs.
We conclude by proposing how the current models could be improved to achieve even higher levels of accuracy using images from hospital scanners.
arXiv Detail & Related papers (2023-10-29T22:31:20Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Efficiently Training Vision Transformers on Structural MRI Scans for
Alzheimer's Disease Detection [2.359557447960552]
Vision transformers (ViT) have emerged in recent years as an alternative to CNNs for several computer vision applications.
We tested variants of the ViT architecture for a range of desired neuroimaging downstream tasks based on difficulty.
We achieved a performance boost of 5% and 9-10% upon fine-tuning vision transformer models pre-trained on synthetic and real MRI scans.
arXiv Detail & Related papers (2023-03-14T20:18:12Z) - AMIGO: Sparse Multi-Modal Graph Transformer with Shared-Context
Processing for Representation Learning of Giga-pixel Images [53.29794593104923]
We present a novel concept of shared-context processing for whole slide histopathology images.
AMIGO uses the celluar graph within the tissue to provide a single representation for a patient.
We show that our model is strongly robust to missing information to an extent that it can achieve the same performance with as low as 20% of the data.
arXiv Detail & Related papers (2023-03-01T23:37:45Z) - CNN-LSTM Based Multimodal MRI and Clinical Data Fusion for Predicting
Functional Outcome in Stroke Patients [1.5250925845050138]
Clinical outcome prediction plays an important role in stroke patient management.
From a machine learning point-of-view, one of the main challenges is dealing with heterogeneous data.
In this paper a multimodal convolutional neural network - long short-term memory (CNN-LSTM) based ensemble model is proposed.
arXiv Detail & Related papers (2022-05-11T14:46:01Z) - A Novel Framework for Brain Tumor Detection Based on Convolutional
Variational Generative Models [6.726255259929498]
This paper introduces a novel framework for brain tumor detection and classification.
The proposed framework acquires an overall detection accuracy of 96.88%.
It highlights the promise of the proposed framework as an accurate low-overhead brain tumor detection system.
arXiv Detail & Related papers (2022-02-20T16:14:01Z) - A self-supervised learning strategy for postoperative brain cavity
segmentation simulating resections [46.414990784180546]
Convolutional neural networks (CNNs) are the state-of-the-art image segmentation technique.
CNNs require large annotated datasets for training.
Self-supervised learning strategies can leverage unlabeled data for training.
arXiv Detail & Related papers (2021-05-24T12:27:06Z) - Transfer Learning Enhanced Generative Adversarial Networks for
Multi-Channel MRI Reconstruction [3.5765797841178597]
Deep learning based generative adversarial networks (GAN) can effectively perform image reconstruction with under-sampled MR data.
It is difficult to obtain tens of thousands of raw patient data to train the model since saving k-space data is not in the routine clinical flow.
In this study, three novel applications were explored based on parallel imaging combined with the GAN model (PI-GAN) and transfer learning.
arXiv Detail & Related papers (2021-05-17T21:28:00Z) - SAG-GAN: Semi-Supervised Attention-Guided GANs for Data Augmentation on
Medical Images [47.35184075381965]
We present a data augmentation method for generating synthetic medical images using cycle-consistency Generative Adversarial Networks (GANs)
The proposed GANs-based model can generate a tumor image from a normal image, and in turn, it can also generate a normal image from a tumor image.
We train the classification model using real images with classic data augmentation methods and classification models using synthetic images.
arXiv Detail & Related papers (2020-11-15T14:01:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.