Integrating Deep Feature Extraction and Hybrid ResNet-DenseNet Model for Multi-Class Abnormality Detection in Endoscopic Images
- URL: http://arxiv.org/abs/2410.18457v1
- Date: Thu, 24 Oct 2024 06:10:31 GMT
- Title: Integrating Deep Feature Extraction and Hybrid ResNet-DenseNet Model for Multi-Class Abnormality Detection in Endoscopic Images
- Authors: Aman Sagar, Preeti Mehta, Monika Shrivastva, Suchi Kumari,
- Abstract summary: The aim is to automate the identification of ten GI abnormality classes, including angioectasia, bleeding, and ulcers.
The proposed model achieves an overall accuracy of 94% across a well-structured dataset.
- Score: 0.9374652839580183
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This paper presents a deep learning framework for the multi-class classification of gastrointestinal abnormalities in Video Capsule Endoscopy (VCE) frames. The aim is to automate the identification of ten GI abnormality classes, including angioectasia, bleeding, and ulcers, thereby reducing the diagnostic burden on gastroenterologists. Utilizing an ensemble of DenseNet and ResNet architectures, the proposed model achieves an overall accuracy of 94\% across a well-structured dataset. Precision scores range from 0.56 for erythema to 1.00 for worms, with recall rates peaking at 98% for normal findings. This study emphasizes the importance of robust data preprocessing techniques, including normalization and augmentation, in enhancing model performance. The contributions of this work lie in developing an effective AI-driven tool that streamlines the diagnostic process in gastroenterology, ultimately improving patient care and clinical outcomes.
Related papers
- FundusGAN: A Hierarchical Feature-Aware Generative Framework for High-Fidelity Fundus Image Generation [35.46876389599076]
FundusGAN is a novel hierarchical feature-aware generative framework specifically designed for high-fidelity fundus image synthesis.
We show that FundusGAN consistently outperforms state-of-the-art methods across multiple metrics.
arXiv Detail & Related papers (2025-03-22T18:08:07Z) - Artificial Intelligence-Driven Prognostic Classification of COVID-19 Using Chest X-rays: A Deep Learning Approach [0.0]
This study presents a high-accuracy deep learning model for classifying COVID-19 severity (Mild, Moderate, and Severe) using Chest X-ray images.
Our model achieved an average accuracy of 97%, with specificity of 99%, sensitivity of 87%, and an F1-score of 93.11%.
These results demonstrate the model's potential for real-world clinical applications.
arXiv Detail & Related papers (2025-03-17T15:27:21Z) - Enhanced Multi-Class Classification of Gastrointestinal Endoscopic Images with Interpretable Deep Learning Model [0.7349657385817541]
This research introduces a novel approach to enhance classification accuracy using 8,000 labeled endoscopic images from the Kvasir dataset.
The proposed architecture eliminates reliance on data augmentation while preserving moderate model complexity.
The model achieves a test accuracy of 94.25%, alongside precision and recall of 94.29% and 94.24% respectively.
arXiv Detail & Related papers (2025-03-02T08:07:50Z) - RURANET++: An Unsupervised Learning Method for Diabetic Macular Edema Based on SCSE Attention Mechanisms and Dynamic Multi-Projection Head Clustering [13.423253964156117]
RURANET++ is an unsupervised learning-based automated diagnostic system for Diabetic Macular Edema (DME)
During feature processing, a pre-trained GoogLeNet model extracts deep features from retinal images, followed by PCA-based dimensionality reduction to 50 dimensions for computational efficiency.
Experimental results demonstrate superior performance across multiple metrics, achieving maximum accuracy (0.8411), precision (0.8593), recall (0.8411), and F1-score, with exceptional clustering quality.
arXiv Detail & Related papers (2025-02-27T16:06:57Z) - EfficientNet with Hybrid Attention Mechanisms for Enhanced Breast Histopathology Classification: A Comprehensive Approach [0.0]
This paper introduces a novel approach integrating Hybrid EfficientNet models with advanced attention mechanisms to enhance feature extraction and focus on critical image regions.
We evaluate the performance of our models across multiple magnification scales using publicly available hispathology datasets.
The results are validated using metrics such as accuracy, F1-score, precision, and recall, demonstrating the clinical potential of our model in improving diagnostic accuracy.
arXiv Detail & Related papers (2024-10-29T17:56:05Z) - CapsuleNet: A Deep Learning Model To Classify GI Diseases Using EfficientNet-b7 [1.2499537119440245]
We present CapsuleNet, a deep learning model developed for the Capsule Vision 2024 Challenge, aimed at classifying 10 distinct GI abnormalities.
Our model leverages a pretrained EfficientNet-b7 backbone, tuned with additional layers for classification and optimized with PReLU activation functions.
Our findings suggest that CNN-based models like CapsuleNet can provide an efficient solution for GI tract disease classification, particularly when inference time is a critical factor.
arXiv Detail & Related papers (2024-10-24T20:43:47Z) - Classification of Endoscopy and Video Capsule Images using CNN-Transformer Model [1.0994755279455526]
This study proposes a hybrid model that combines the advantages of Transformers and Convolutional Neural Networks (CNNs) to enhance classification performance.
For the GastroVision dataset, our proposed model demonstrates excellent performance with Precision, Recall, F1 score, Accuracy, and Matthews Correlation Coefficient (MCC) of 0.8320, 0.8386, 0.8324, 0.8386, and 0.8191, respectively.
arXiv Detail & Related papers (2024-08-20T11:05:32Z) - Enhancing Diabetic Retinopathy Diagnosis: A Lightweight CNN Architecture for Efficient Exudate Detection in Retinal Fundus Images [0.0]
This paper introduces a novel, lightweight convolutional neural network architecture tailored for automated exudate detection.
We have incorporated domain-specific data augmentations to enhance the model's generalizability.
Our model achieves an impressive F1 score of 90%, demonstrating its efficacy in the early detection of diabetic retinopathy through fundus imaging.
arXiv Detail & Related papers (2024-08-13T10:13:33Z) - Benchmarking Embedding Aggregation Methods in Computational Pathology: A Clinical Data Perspective [32.93871326428446]
Recent advances in artificial intelligence (AI) are revolutionizing medical imaging and computational pathology.
A constant challenge in the analysis of digital Whole Slide Images (WSIs) is the problem of aggregating tens of thousands of tile-level image embeddings to a slide-level representation.
This study conducts a benchmarking analysis of ten slide-level aggregation techniques across nine clinically relevant tasks.
arXiv Detail & Related papers (2024-07-10T17:00:57Z) - Deep learning in computed tomography pulmonary angiography imaging: a
dual-pronged approach for pulmonary embolism detection [0.0]
The aim of this study is to leverage deep learning techniques to enhance the Computer Assisted Diagnosis (CAD) of Pulmonary Embolism (PE)
Our classification system includes an Attention-Guided Convolutional Neural Network (AG-CNN) that uses local context by employing an attention mechanism.
AG-CNN achieves robust performance on the FUMPE dataset, achieving an AUROC of 0.927, sensitivity of 0.862, specificity of 0.879, and an F1-score of 0.805 with the Inception-v3 backbone architecture.
arXiv Detail & Related papers (2023-11-09T08:23:44Z) - Neural Network-Based Histologic Remission Prediction In Ulcerative
Colitis [38.150634108667774]
Histologic remission is a new therapeutic target in ulcerative colitis (UC)
Endocytoscopy (EC) is a novel ultra-high magnification endoscopic technique.
We propose a neural network model that can assess histological disease activity in EC images.
arXiv Detail & Related papers (2023-08-28T15:54:14Z) - Automatic diagnosis of knee osteoarthritis severity using Swin
transformer [55.01037422579516]
Knee osteoarthritis (KOA) is a widespread condition that can cause chronic pain and stiffness in the knee joint.
We propose an automated approach that employs the Swin Transformer to predict the severity of KOA.
arXiv Detail & Related papers (2023-07-10T09:49:30Z) - CIRCA: comprehensible online system in support of chest X-rays-based
COVID-19 diagnosis [37.41181188499616]
Deep learning techniques can help in the faster detection of COVID-19 cases and monitoring of disease progression.
Five different datasets were used to construct a representative dataset of 23 799 CXRs for model training.
A U-Net-based model was developed to identify a clinically relevant region of the CXR.
arXiv Detail & Related papers (2022-10-11T13:30:34Z) - HistoPerm: A Permutation-Based View Generation Approach for Improving
Histopathologic Feature Representation Learning [33.1098457952173]
HistoPerm is a view generation method for representation learning using joint embedding architectures.
HistoPerm permutes augmented views of patches extracted from whole-slide histology images to improve classification performance.
Our results show that HistoPerm consistently improves patch- and slide-level classification performance in terms of accuracy, F1-score, and AUC.
arXiv Detail & Related papers (2022-09-13T17:35:08Z) - Global ECG Classification by Self-Operational Neural Networks with
Feature Injection [25.15075119957447]
We propose a novel approach for inter-patient ECG classification using a compact 1D Self-Organized Operational Neural Networks (Self-ONNs)
We used 1D Self-ONN layers to automatically learn morphological representations from ECG data, enabling us to capture the shape of the ECG waveform around the R peaks.
Using the MIT-BIH arrhythmia benchmark database, the proposed method achieves the highest classification performance ever achieved.
arXiv Detail & Related papers (2022-04-07T22:49:18Z) - Multi-Task Neural Networks with Spatial Activation for Retinal Vessel
Segmentation and Artery/Vein Classification [49.64863177155927]
We propose a multi-task deep neural network with spatial activation mechanism to segment full retinal vessel, artery and vein simultaneously.
The proposed network achieves pixel-wise accuracy of 95.70% for vessel segmentation, and A/V classification accuracy of 94.50%, which is the state-of-the-art performance for both tasks.
arXiv Detail & Related papers (2020-07-18T05:46:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.