A ResNet is All You Need? Modeling A Strong Baseline for Detecting
Referable Diabetic Retinopathy in Fundus Images
- URL: http://arxiv.org/abs/2210.03180v1
- Date: Thu, 6 Oct 2022 19:40:56 GMT
- Title: A ResNet is All You Need? Modeling A Strong Baseline for Detecting
Referable Diabetic Retinopathy in Fundus Images
- Authors: Tom\'as Castilla, Marcela S. Mart\'inez, Mercedes Legu\'ia, Ignacio
Larrabide, Jos\'e Ignacio Orlando
- Abstract summary: We model a strong baseline for this task based on a simple and standard ResNet-18 architecture.
Our model achieved an AUC = 0.955 on a combined test set of 61007 test images from different public datasets.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning is currently the state-of-the-art for automated detection of
referable diabetic retinopathy (DR) from color fundus photographs (CFP). While
the general interest is put on improving results through methodological
innovations, it is not clear how good these approaches perform compared to
standard deep classification models trained with the appropriate settings. In
this paper we propose to model a strong baseline for this task based on a
simple and standard ResNet-18 architecture. To this end, we built on top of
prior art by training the model with a standard preprocessing strategy but
using images from several public sources and an empirically calibrated data
augmentation setting. To evaluate its performance, we covered multiple
clinically relevant perspectives, including image and patient level DR
screening, discriminating responses by input quality and DR grade, assessing
model uncertainties and analyzing its results in a qualitative manner. With no
other methodological innovation than a carefully designed training, our ResNet
model achieved an AUC = 0.955 (0.953 - 0.956) on a combined test set of 61007
test images from different public datasets, which is in line or even better
than what other more complex deep learning models reported in the literature.
Similar AUC values were obtained in 480 images from two separate in-house
databases specially prepared for this study, which emphasize its generalization
ability. This confirms that standard networks can still be strong baselines for
this task if properly trained.
Related papers
- Reinforcing Pre-trained Models Using Counterfactual Images [54.26310919385808]
This paper proposes a novel framework to reinforce classification models using language-guided generated counterfactual images.
We identify model weaknesses by testing the model using the counterfactual image dataset.
We employ the counterfactual images as an augmented dataset to fine-tune and reinforce the classification model.
arXiv Detail & Related papers (2024-06-19T08:07:14Z) - Opinion-Unaware Blind Image Quality Assessment using Multi-Scale Deep Feature Statistics [54.08757792080732]
We propose integrating deep features from pre-trained visual models with a statistical analysis model to achieve opinion-unaware BIQA (OU-BIQA)
Our proposed model exhibits superior consistency with human visual perception compared to state-of-the-art BIQA models.
arXiv Detail & Related papers (2024-05-29T06:09:34Z) - Uncovering the effects of model initialization on deep model
generalization: A study with adult and pediatric Chest X-ray images [5.454938535500864]
ImageNet-pretrained weights demonstrate superior generalizability over randomly counterparts, contradicting some findings for non-medical images.
Weight-level ensembles of these models show significantly higher recall (p 0.05) during testing compared to individual models.
arXiv Detail & Related papers (2023-09-20T13:42:48Z) - MedFMC: A Real-world Dataset and Benchmark For Foundation Model
Adaptation in Medical Image Classification [41.16626194300303]
Foundation models, often pre-trained with large-scale data, have achieved paramount success in jump-starting various vision and language applications.
Recent advances further enable adapting foundation models in downstream tasks efficiently using only a few training samples.
Yet, the application of such learning paradigms in medical image analysis remains scarce due to the shortage of publicly accessible data and benchmarks.
arXiv Detail & Related papers (2023-06-16T01:46:07Z) - Universal Domain Adaptation from Foundation Models: A Baseline Study [58.51162198585434]
We make empirical studies of state-of-the-art UniDA methods using foundation models.
We introduce textitCLIP distillation, a parameter-free method specifically designed to distill target knowledge from CLIP models.
Although simple, our method outperforms previous approaches in most benchmark tasks.
arXiv Detail & Related papers (2023-05-18T16:28:29Z) - Self-similarity-based super-resolution of photoacoustic angiography from
hand-drawn doodles [9.646172419644088]
Photoacoustic angiography (PAA) is a powerful tool that restores blood vessel images from under-sampled images to facilitate disease diagnosis.
We propose a novel approach that employs a super-resolution PAA method trained with forged PAA images.
Our approach shows promising potential for exploring zero-shot learning neural networks for vision tasks.
arXiv Detail & Related papers (2023-05-02T02:40:47Z) - Vision-Language Modelling For Radiological Imaging and Reports In The
Low Data Regime [70.04389979779195]
This paper explores training medical vision-language models (VLMs) where the visual and language inputs are embedded into a common space.
We explore several candidate methods to improve low-data performance, including adapting generic pre-trained models to novel image and text domains.
Using text-to-image retrieval as a benchmark, we evaluate the performance of these methods with variable sized training datasets of paired chest X-rays and radiological reports.
arXiv Detail & Related papers (2023-03-30T18:20:00Z) - An Ensemble Method to Automatically Grade Diabetic Retinopathy with
Optical Coherence Tomography Angiography Images [4.640835690336653]
We propose an ensemble method to automatically grade Diabetic retinopathy (DR) images available from Diabetic Retinopathy Analysis Challenge (DRAC) 2022.
First, we adopt the state-of-the-art classification networks, and train them to grade UW- OCTA images with different splits of the available dataset.
Ultimately, we obtain 25 models, of which, the top 16 models are selected and ensembled to generate the final predictions.
arXiv Detail & Related papers (2022-12-12T22:06:47Z) - FundusQ-Net: a Regression Quality Assessment Deep Learning Algorithm for
Fundus Images Quality Grading [0.0]
Glaucoma, diabetic retinopathy and age-related macular degeneration are major causes of blindness and vision impairment.
Key step in this process is to automatically estimate the quality of the fundus images to make sure these are interpretable by a human operator or a machine learning model.
We present a novel fundus image quality scale and deep learning (DL) model that can estimate fundus image quality relative to this new scale.
arXiv Detail & Related papers (2022-05-02T21:01:34Z) - On the Robustness of Pretraining and Self-Supervision for a Deep
Learning-based Analysis of Diabetic Retinopathy [70.71457102672545]
We compare the impact of different training procedures for diabetic retinopathy grading.
We investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions.
Our results indicate that models from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions.
arXiv Detail & Related papers (2021-06-25T08:32:45Z) - Improved Techniques for Training Score-Based Generative Models [104.20217659157701]
We provide a new theoretical analysis of learning and sampling from score models in high dimensional spaces.
We can effortlessly scale score-based generative models to images with unprecedented resolutions.
Our score-based models can generate high-fidelity samples that rival best-in-class GANs on various image datasets.
arXiv Detail & Related papers (2020-06-16T09:17:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.