Blended Multi-Modal Deep ConvNet Features for Diabetic Retinopathy
Severity Prediction
- URL: http://arxiv.org/abs/2006.00197v1
- Date: Sat, 30 May 2020 06:46:26 GMT
- Title: Blended Multi-Modal Deep ConvNet Features for Diabetic Retinopathy
Severity Prediction
- Authors: J.D. Bodapati, N. Veeranjaneyulu, S.N. Shareef, S. Hakak, M. Bilal,
P.K.R. Maddikunta, O. Jo
- Abstract summary: Diabetic Retinopathy (DR) is one of the major causes of visual impairment and blindness across the world.
To derive optimal representation of retinal images, features extracted from multiple pre-trained ConvNet models are blended.
We achieve an accuracy of 97.41%, and a kappa statistic of 94.82 for DR identification and an accuracy of 81.7% and a kappa statistic of 71.1% for severity level prediction.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Diabetic Retinopathy (DR) is one of the major causes of visual impairment and
blindness across the world. It is usually found in patients who suffer from
diabetes for a long period. The major focus of this work is to derive optimal
representation of retinal images that further helps to improve the performance
of DR recognition models. To extract optimal representation, features extracted
from multiple pre-trained ConvNet models are blended using proposed multi-modal
fusion module. These final representations are used to train a Deep Neural
Network (DNN) used for DR identification and severity level prediction. As each
ConvNet extracts different features, fusing them using 1D pooling and cross
pooling leads to better representation than using features extracted from a
single ConvNet. Experimental studies on benchmark Kaggle APTOS 2019 contest
dataset reveals that the model trained on proposed blended feature
representations is superior to the existing methods. In addition, we notice
that cross average pooling based fusion of features from Xception and VGG16 is
the most appropriate for DR recognition. With the proposed model, we achieve an
accuracy of 97.41%, and a kappa statistic of 94.82 for DR identification and an
accuracy of 81.7% and a kappa statistic of 71.1% for severity level prediction.
Another interesting observation is that DNN with dropout at input layer
converges more quickly when trained using blended features, compared to the
same model trained using uni-modal deep features.
Related papers
- Deep Learning Ensemble for Predicting Diabetic Macular Edema Onset Using Ultra-Wide Field Color Fundus Image [3.271278111396875]
Diabetic macular edema (DME) is a severe complication of diabetes, characterized by thickening of the central portion of the retina due to accumulation of fluid.
We propose an ensemble method to predict ci-DME onset within a year using ultra-wide-field color fundus photography images.
arXiv Detail & Related papers (2024-10-09T02:16:29Z) - Brain Tumor Classification on MRI in Light of Molecular Markers [61.77272414423481]
Co-deletion of the 1p/19q gene is associated with clinical outcomes in low-grade gliomas.
This study aims to utilize a specially MRI-based convolutional neural network for brain cancer detection.
arXiv Detail & Related papers (2024-09-29T07:04:26Z) - Comparative Performance Analysis of Transformer-Based Pre-Trained Models for Detecting Keratoconus Disease [0.0]
This study compares eight pre-trained CNNs for diagnosing keratoconus, a degenerative eye disease.
MobileNetV2 was the best accurate model in identifying keratoconus and normal cases with few misclassifications.
arXiv Detail & Related papers (2024-08-16T20:15:24Z) - Cross-Model Comparative Loss for Enhancing Neuronal Utility in Language
Understanding [82.46024259137823]
We propose a cross-model comparative loss for a broad range of tasks.
We demonstrate the universal effectiveness of comparative loss through extensive experiments on 14 datasets from 3 distinct NLU tasks.
arXiv Detail & Related papers (2023-01-10T03:04:27Z) - An Ensemble Method to Automatically Grade Diabetic Retinopathy with
Optical Coherence Tomography Angiography Images [4.640835690336653]
We propose an ensemble method to automatically grade Diabetic retinopathy (DR) images available from Diabetic Retinopathy Analysis Challenge (DRAC) 2022.
First, we adopt the state-of-the-art classification networks, and train them to grade UW- OCTA images with different splits of the available dataset.
Ultimately, we obtain 25 models, of which, the top 16 models are selected and ensembled to generate the final predictions.
arXiv Detail & Related papers (2022-12-12T22:06:47Z) - CNN-based fully automatic wrist cartilage volume quantification in MR
Image [55.41644538483948]
The U-net convolutional neural network with additional attention layers provides the best wrist cartilage segmentation performance.
The error of cartilage volume measurement should be assessed independently using a non-MRI method.
arXiv Detail & Related papers (2022-06-22T14:19:06Z) - FEDI: Few-shot learning based on Earth Mover's Distance algorithm
combined with deep residual network to identify diabetic retinopathy [3.6623193507510012]
This paper proposes a few-shot learning model of a deep residual network based on Earth Mover's algorithm to assist in diagnosing diabetic retinopathy.
We build training and validation classification tasks for few-shot learning based on 39 categories of 1000 sample data, train deep residual networks, and obtain experience pre-training models.
Based on the weights of the pre-trained model, the Earth Mover's Distance algorithm calculates the distance between the images, obtains the similarity between the images, and changes the model's parameters to improve the accuracy of the training model.
arXiv Detail & Related papers (2021-08-22T13:05:02Z) - Vision Transformers for femur fracture classification [59.99241204074268]
The Vision Transformer (ViT) was able to correctly predict 83% of the test images.
Good results were obtained in sub-fractures with the largest and richest dataset ever.
arXiv Detail & Related papers (2021-08-07T10:12:42Z) - Predictive Analysis of Diabetic Retinopathy with Transfer Learning [0.0]
This paper studies the performance of CNN architectures for Diabetic Retinopathy Classification with the help of Transfer Learning.
The results indicate that Transfer Learning with ImageNet weights using VGG 16 model demonstrates the best classification performance with the best Accuracy of 95%.
arXiv Detail & Related papers (2020-11-08T18:54:57Z) - Classification of COVID-19 in CT Scans using Multi-Source Transfer
Learning [91.3755431537592]
We propose the use of Multi-Source Transfer Learning to improve upon traditional Transfer Learning for the classification of COVID-19 from CT scans.
With our multi-source fine-tuning approach, our models outperformed baseline models fine-tuned with ImageNet.
Our best performing model was able to achieve an accuracy of 0.893 and a Recall score of 0.897, outperforming its baseline Recall score by 9.3%.
arXiv Detail & Related papers (2020-09-22T11:53:06Z) - M2Net: Multi-modal Multi-channel Network for Overall Survival Time
Prediction of Brain Tumor Patients [151.4352001822956]
Early and accurate prediction of overall survival (OS) time can help to obtain better treatment planning for brain tumor patients.
Existing prediction methods rely on radiomic features at the local lesion area of a magnetic resonance (MR) volume.
We propose an end-to-end OS time prediction model; namely, Multi-modal Multi-channel Network (M2Net)
arXiv Detail & Related papers (2020-06-01T05:21:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.