Capsule Endoscopy Multi-classification via Gated Attention and Wavelet Transformations
- URL: http://arxiv.org/abs/2410.19363v1
- Date: Fri, 25 Oct 2024 08:01:35 GMT
- Title: Capsule Endoscopy Multi-classification via Gated Attention and Wavelet Transformations
- Authors: Lakshmi Srinivas Panchananam, Praveen Kumar Chandaliya, Kishor Upla, Kiran Raja,
- Abstract summary: Abnormalities in the gastrointestinal tract significantly influence the patient's health and require a timely diagnosis.
The work presents the process of developing and evaluating a novel model designed to classify gastrointestinal anomalies from a video frame.
integration of Omni Dimensional Gated Attention (OGA) mechanism and Wavelet transformation techniques into the model's architecture allowed the model to focus on the most critical areas.
The model's performance is benchmarked against two base models, VGG16 and ResNet50, demonstrating its enhanced ability to identify and classify a range of gastrointestinal abnormalities accurately.
- Score: 1.5146068448101746
- License:
- Abstract: Abnormalities in the gastrointestinal tract significantly influence the patient's health and require a timely diagnosis for effective treatment. With such consideration, an effective automatic classification of these abnormalities from a video capsule endoscopy (VCE) frame is crucial for improvement in diagnostic workflows. The work presents the process of developing and evaluating a novel model designed to classify gastrointestinal anomalies from a VCE video frame. Integration of Omni Dimensional Gated Attention (OGA) mechanism and Wavelet transformation techniques into the model's architecture allowed the model to focus on the most critical areas in the endoscopy images, reducing noise and irrelevant features. This is particularly advantageous in capsule endoscopy, where images often contain a high degree of variability in texture and color. Wavelet transformations contributed by efficiently capturing spatial and frequency-domain information, improving feature extraction, especially for detecting subtle features from the VCE frames. Furthermore, the features extracted from the Stationary Wavelet Transform and Discrete Wavelet Transform are concatenated channel-wise to capture multiscale features, which are essential for detecting polyps, ulcerations, and bleeding. This approach improves classification accuracy on imbalanced capsule endoscopy datasets. The proposed model achieved 92.76% and 91.19% as training and validation accuracies respectively. At the same time, Training and Validation losses are 0.2057 and 0.2700. The proposed model achieved a Balanced Accuracy of 94.81%, AUC of 87.49%, F1-score of 91.11%, precision of 91.17%, recall of 91.19% and specificity of 98.44%. Additionally, the model's performance is benchmarked against two base models, VGG16 and ResNet50, demonstrating its enhanced ability to identify and classify a range of gastrointestinal abnormalities accurately.
Related papers
- Integrating Deep Feature Extraction and Hybrid ResNet-DenseNet Model for Multi-Class Abnormality Detection in Endoscopic Images [0.9374652839580183]
The aim is to automate the identification of ten GI abnormality classes, including angioectasia, bleeding, and ulcers.
The proposed model achieves an overall accuracy of 94% across a well-structured dataset.
arXiv Detail & Related papers (2024-10-24T06:10:31Z) - Controllable retinal image synthesis using conditional StyleGAN and latent space manipulation for improved diagnosis and grading of diabetic retinopathy [0.0]
This paper proposes a framework for controllably generating high-fidelity and diverse DR fundus images.
We achieve comprehensive control over DR severity and visual features within generated images.
We manipulate the DR images generated conditionally on grades, further enhancing the dataset diversity.
arXiv Detail & Related papers (2024-09-11T17:08:28Z) - Deep Generative Classification of Blood Cell Morphology [7.494975467007647]
We introduce CytoDiffusion, a diffusion-based classifier that effectively models blood cell morphology.
Our approach outperforms state-of-the-art discriminative models in anomaly detection.
We enhance model explainability through the generation of directly interpretable counterfactual heatmaps.
arXiv Detail & Related papers (2024-08-16T19:17:02Z) - Efficient Data-Sketches and Fine-Tuning for Early Detection of Distributional Drift in Medical Imaging [5.1358645354733765]
This paper presents an accurate and sensitive approach to detect distributional drift in CT-scan medical images.
We developed a robust library model for real-time anomaly detection, allowing for efficient comparison of incoming images.
We fine-tuned a vision transformer pre-trained model to extract relevant features using breast cancer images.
arXiv Detail & Related papers (2024-08-15T23:46:37Z) - Enhancing Diagnostic Reliability of Foundation Model with Uncertainty Estimation in OCT Images [41.002573031087856]
We developed a foundation model with uncertainty estimation (FMUE) to detect 11 retinal conditions on optical coherence tomography ( OCT)
FMUE achieved a higher F1 score of 96.76% than two state-of-the-art algorithms, RETFound and UIOS, and got further improvement with thresholding strategy to 98.44%.
Our model is superior to two ophthalmologists with a higher F1 score (95.17% vs. 61.93% &71.72%)
arXiv Detail & Related papers (2024-06-18T03:04:52Z) - Detection of Peri-Pancreatic Edema using Deep Learning and Radiomics Techniques [3.6663216419047964]
This study textitintroduces a novel CT dataset sourced from 255 patients with pancreatic diseases, featuring annotated pancreas segmentation masks and corresponding diagnostic labels for peri-pancreatic edema condition.
We first evaluate the efficacy of the textitLinTransUNet model, a linear Transformer based segmentation algorithm, to segment the pancreas accurately from CT imaging data.
Then, we use segmented pancreas regions with two distinctive machine learning classifiers to identify existence of peri-pancreatic edema: deep learning-based models and a radiomics-based eXtreme Gradient Boosting (XGBoost)
arXiv Detail & Related papers (2024-04-25T22:20:17Z) - Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - Texture Characterization of Histopathologic Images Using Ecological
Diversity Measures and Discrete Wavelet Transform [82.53597363161228]
This paper proposes a method for characterizing texture across histopathologic images with a considerable success rate.
It is possible to quantify the intrinsic properties of such images with promising accuracy on two HI datasets.
arXiv Detail & Related papers (2022-02-27T02:19:09Z) - A multicenter study on radiomic features from T$_2$-weighted images of a
customized MR pelvic phantom setting the basis for robust radiomic models in
clinics [47.187609203210705]
2D and 3D T$$-weighted images of a pelvic phantom were acquired on three scanners.
repeatability and repositioning of radiomic features were assessed.
arXiv Detail & Related papers (2020-05-14T09:24:48Z) - Adaptive Feature Selection Guided Deep Forest for COVID-19
Classification with Chest CT [49.09507792800059]
We propose an Adaptive Feature Selection guided Deep Forest (AFS-DF) for COVID-19 classification based on chest CT images.
We evaluate our proposed AFS-DF on COVID-19 dataset with 1495 patients of COVID-19 and 1027 patients of community acquired pneumonia (CAP)
arXiv Detail & Related papers (2020-05-07T06:00:02Z) - Residual Attention U-Net for Automated Multi-Class Segmentation of
COVID-19 Chest CT Images [46.844349956057776]
coronavirus disease 2019 (COVID-19) has been spreading rapidly around the world and caused significant impact on the public health and economy.
There is still lack of studies on effectively quantifying the lung infection caused by COVID-19.
We propose a novel deep learning algorithm for automated segmentation of multiple COVID-19 infection regions.
arXiv Detail & Related papers (2020-04-12T16:24:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.