Gastrointestinal Disorder Detection with a Transformer Based Approach
- URL: http://arxiv.org/abs/2210.03168v1
- Date: Thu, 6 Oct 2022 19:08:37 GMT
- Title: Gastrointestinal Disorder Detection with a Transformer Based Approach
- Authors: A.K.M. Salman Hosain, Mynul islam, Md Humaion Kabir Mehedi, Irteza
Enan Kabir, Zarin Tasnim Khan
- Abstract summary: This paper describes a technique for assisting medical diagnosis procedures and identifying gastrointestinal tract disorders based on the categorization of characteristics taken from endoscopic pictures.
We have suggested a vision transformer based approach to detect gastrointestianl diseases from wireless capsule endoscopy (WCE) curated images of colon with an accuracy of 95.63%.
We have compared this transformer based approach with pretrained convolutional neural network (CNN) model DenseNet201 and demonstrated that vision transformer surpassed DenseNet201 in various quantitative performance evaluation metrics.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate disease categorization using endoscopic images is a significant
problem in Gastroenterology. This paper describes a technique for assisting
medical diagnosis procedures and identifying gastrointestinal tract disorders
based on the categorization of characteristics taken from endoscopic pictures
using a vision transformer and transfer learning model. Vision transformer has
shown very promising results on difficult image classification tasks. In this
paper, we have suggested a vision transformer based approach to detect
gastrointestianl diseases from wireless capsule endoscopy (WCE) curated images
of colon with an accuracy of 95.63\%. We have compared this transformer based
approach with pretrained convolutional neural network (CNN) model DenseNet201
and demonstrated that vision transformer surpassed DenseNet201 in various
quantitative performance evaluation metrics.
Related papers
- Classification of Endoscopy and Video Capsule Images using CNN-Transformer Model [1.0994755279455526]
This study proposes a hybrid model that combines the advantages of Transformers and Convolutional Neural Networks (CNNs) to enhance classification performance.
For the GastroVision dataset, our proposed model demonstrates excellent performance with Precision, Recall, F1 score, Accuracy, and Matthews Correlation Coefficient (MCC) of 0.8320, 0.8386, 0.8324, 0.8386, and 0.8191, respectively.
arXiv Detail & Related papers (2024-08-20T11:05:32Z) - Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - Breast Ultrasound Tumor Classification Using a Hybrid Multitask
CNN-Transformer Network [63.845552349914186]
Capturing global contextual information plays a critical role in breast ultrasound (BUS) image classification.
Vision Transformers have an improved capability of capturing global contextual information but may distort the local image patterns due to the tokenization operations.
In this study, we proposed a hybrid multitask deep neural network called Hybrid-MT-ESTAN, designed to perform BUS tumor classification and segmentation.
arXiv Detail & Related papers (2023-08-04T01:19:32Z) - Reliable Joint Segmentation of Retinal Edema Lesions in OCT Images [55.83984261827332]
In this paper, we propose a novel reliable multi-scale wavelet-enhanced transformer network.
We develop a novel segmentation backbone that integrates a wavelet-enhanced feature extractor network and a multi-scale transformer module.
Our proposed method achieves better segmentation accuracy with a high degree of reliability as compared to other state-of-the-art segmentation approaches.
arXiv Detail & Related papers (2022-12-01T07:32:56Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - Shifted Windows Transformers for Medical Image Quality Assessment [0.0]
CNN-based approaches are used to assess the image quality, but their performance can still be improved in terms of accuracy.
In this work, we approach this problem by using Swin Transformer, which improves the poor-quality image classification performance.
To the best of our knowledge, our study is the first vision transformer application for medical image quality assessment.
arXiv Detail & Related papers (2022-08-11T20:39:26Z) - Gastrointestinal Polyps and Tumors Detection Based on Multi-scale
Feature-fusion with WCE Sequences [0.0]
This paper proposes a textbfTwo-stage textbfMulti-scale textbfFeature-fusion learning network(textbfTMFNet) to automatically detect small intestinal polyps and tumors.
We used 22,335 WCE images in the experiment, with a total of 123,092 lesion regions used to train the detection framework of this paper.
arXiv Detail & Related papers (2022-04-03T07:24:50Z) - CyTran: A Cycle-Consistent Transformer with Multi-Level Consistency for
Non-Contrast to Contrast CT Translation [56.622832383316215]
We propose a novel approach to translate unpaired contrast computed tomography (CT) scans to non-contrast CT scans.
Our approach is based on cycle-consistent generative adversarial convolutional transformers, for short, CyTran.
Our empirical results show that CyTran outperforms all competing methods.
arXiv Detail & Related papers (2021-10-12T23:25:03Z) - GasHis-Transformer: A Multi-scale Visual Transformer Approach for
Gastric Histopathology Image Classification [30.497184157710873]
This paper proposes a multi-scale visual transformer model (GasHis-Transformer) for a gastric histopathology image classification (GHIC) task.
GasHis-Transformer model is built on two fundamental modules, including a global information module (GIM) and a local information module (LIM)
arXiv Detail & Related papers (2021-04-29T17:46:00Z) - Pyramid Focusing Network for mutation prediction and classification in
CT images [2.4440097656693553]
We propose a pyramid focusing network (PFNet) for mutation prediction and classification based on CT images.
Our method achieves the accuracy of 94.90% in predicting the HER-2 genes mutation status at the CT image.
arXiv Detail & Related papers (2020-04-07T12:14:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.