Skin Cancer Segmentation and Classification Using Vision Transformer for
Automatic Analysis in Dermatoscopy-based Non-invasive Digital System
- URL: http://arxiv.org/abs/2401.04746v1
- Date: Tue, 9 Jan 2024 11:22:54 GMT
- Title: Skin Cancer Segmentation and Classification Using Vision Transformer for
Automatic Analysis in Dermatoscopy-based Non-invasive Digital System
- Authors: Galib Muhammad Shahriar Himel, Md. Masudul Islam, Kh Abdullah Al-Aff,
Shams Ibne Karim, Md. Kabir Uddin Sikder
- Abstract summary: This study introduces a groundbreaking approach to skin cancer classification, employing the Vision Transformer.
The Vision Transformer is a state-of-the-art deep learning architecture renowned for its success in diverse image analysis tasks.
The Segment Anything Model aids in precise segmentation of cancerous areas, attaining high IOU and Dice Coefficient.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Skin cancer is a global health concern, necessitating early and accurate
diagnosis for improved patient outcomes. This study introduces a groundbreaking
approach to skin cancer classification, employing the Vision Transformer, a
state-of-the-art deep learning architecture renowned for its success in diverse
image analysis tasks. Utilizing the HAM10000 dataset of 10,015 meticulously
annotated skin lesion images, the model undergoes preprocessing for enhanced
robustness. The Vision Transformer, adapted to the skin cancer classification
task, leverages the self-attention mechanism to capture intricate spatial
dependencies, achieving superior performance over traditional deep learning
architectures. Segment Anything Model aids in precise segmentation of cancerous
areas, attaining high IOU and Dice Coefficient. Extensive experiments highlight
the model's supremacy, particularly the Google-based ViT patch-32 variant,
which achieves 96.15% accuracy and showcases potential as an effective tool for
dermatologists in skin cancer diagnosis, contributing to advancements in
dermatological practices.
Related papers
- Advanced Hybrid Deep Learning Model for Enhanced Classification of Osteosarcoma Histopathology Images [0.0]
This study focuses on osteosarcoma (OS), the most common bone cancer in children and adolescents, which affects the long bones of the arms and legs.
We propose a novel hybrid model that combines convolutional neural networks (CNN) and vision transformers (ViT) to improve diagnostic accuracy for OS.
The model achieved an accuracy of 99.08%, precision of 99.10%, recall of 99.28%, and an F1-score of 99.23%.
arXiv Detail & Related papers (2024-10-29T13:54:08Z) - Comparative Analysis of Transfer Learning Models for Breast Cancer Classification [10.677937909900486]
This study investigates the efficiency of deep learning models in distinguishing between Invasive Ductal Carcinoma (IDC) and non-IDC in histopathology slides.
We conducted a comparison examination of eight sophisticated models: ResNet-50, DenseNet-121, ResNeXt-50, Vision Transformer (ViT), GoogLeNet (Inception v3), EfficientNet, MobileNet, and SqueezeNet.
arXiv Detail & Related papers (2024-08-29T18:49:32Z) - Hybrid Deep Learning Framework for Enhanced Melanoma Detection [3.004788114489393]
The primary objective of our study is to enhance the accuracy and efficiency of melanoma detection through an innovative hybrid approach.
We utilize the HAM10000 dataset to meticulously train the U-Net model, enabling it to precisely segment cancerous regions.
We employ the ISIC 2020 dataset to train the EfficientNet model, optimizing it for the binary classification of skin cancer.
arXiv Detail & Related papers (2024-07-16T04:58:47Z) - Optimizing Synthetic Correlated Diffusion Imaging for Breast Cancer Tumour Delineation [71.91773485443125]
We show that the best AUC is achieved by the CDI$s$ - optimized modality, outperforming the best gold-standard modality by 0.0044.
Notably, the optimized CDI$s$ modality also achieves AUC values over 0.02 higher than the Unoptimized CDI$s$ value.
arXiv Detail & Related papers (2024-05-13T16:07:58Z) - Optimizing Skin Lesion Classification via Multimodal Data and Auxiliary
Task Integration [54.76511683427566]
This research introduces a novel multimodal method for classifying skin lesions, integrating smartphone-captured images with essential clinical and demographic information.
A distinctive aspect of this method is the integration of an auxiliary task focused on super-resolution image prediction.
The experimental evaluations have been conducted using the PAD-UFES20 dataset, applying various deep-learning architectures.
arXiv Detail & Related papers (2024-02-16T05:16:20Z) - Multi-Scale Hybrid Vision Transformer for Learning Gastric Histology:
AI-Based Decision Support System for Gastric Cancer Treatment [50.89811515036067]
Gastric endoscopic screening is an effective way to decide appropriate gastric cancer (GC) treatment at an early stage, reducing GC-associated mortality rate.
We propose a practical AI system that enables five subclassifications of GC pathology, which can be directly matched to general GC treatment guidance.
arXiv Detail & Related papers (2022-02-17T08:33:52Z) - DenseNet approach to segmentation and classification of dermatoscopic
skin lesions images [0.0]
This paper proposes an improved method for segmentation and classification for skin lesions using two architectures.
The combination of U-Net and DenseNet121 provides acceptable results in dermatoscopic image analysis.
cancerous and non-cancerous samples were detected in DenseNet121 network with 79.49% and 93.11% accuracy respectively.
arXiv Detail & Related papers (2021-10-09T19:12:23Z) - OncoPetNet: A Deep Learning based AI system for mitotic figure counting
on H&E stained whole slide digital images in a large veterinary diagnostic
lab setting [47.38796928990688]
Multiple state-of-the-art deep learning techniques for histopathology image classification and mitotic figure detection were used in the development of OncoPetNet.
The proposed system, demonstrated significantly improved mitotic counting performance for 41 cancer cases across 14 cancer types compared to human expert baselines.
In deployment, an effective 0.27 min/slide inference was achieved in a high throughput veterinary diagnostic service across 2 centers processing 3,323 digital whole slide images daily.
arXiv Detail & Related papers (2021-08-17T20:01:33Z) - Spectral-Spatial Recurrent-Convolutional Networks for In-Vivo
Hyperspectral Tumor Type Classification [49.32653090178743]
We demonstrate the feasibility of in-vivo tumor type classification using hyperspectral imaging and deep learning.
Our best model achieves an AUC of 76.3%, significantly outperforming previous conventional and deep learning methods.
arXiv Detail & Related papers (2020-07-02T12:00:53Z) - Harvesting, Detecting, and Characterizing Liver Lesions from Large-scale
Multi-phase CT Data via Deep Dynamic Texture Learning [24.633802585888812]
We propose a fully-automated and multi-stage liver tumor characterization framework for dynamic contrast computed tomography (CT)
Our system comprises four sequential processes of tumor proposal detection, tumor harvesting, primary tumor site selection, and deep texture-based tumor characterization.
arXiv Detail & Related papers (2020-06-28T19:55:34Z) - Spatio-spectral deep learning methods for in-vivo hyperspectral
laryngeal cancer detection [49.32653090178743]
Early detection of head and neck tumors is crucial for patient survival.
Hyperspectral imaging (HSI) can be used for non-invasive detection of head and neck tumors.
We present multiple deep learning techniques for in-vivo laryngeal cancer detection based on HSI.
arXiv Detail & Related papers (2020-04-21T17:07:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.