Learning from few examples: Classifying sex from retinal images via deep
learning
- URL: http://arxiv.org/abs/2207.09624v1
- Date: Wed, 20 Jul 2022 02:47:29 GMT
- Title: Learning from few examples: Classifying sex from retinal images via deep
learning
- Authors: Aaron Berk, Gulcenur Ozturan, Parsa Delavari, David Maberley,
\"Ozg\"ur Y{\i}lmaz and Ipek Oruc
- Abstract summary: We showcase results for the performance of DL on small datasets to classify patient sex from fundus images.
Our models, developed using approximately 2500 fundus images, achieved test AUC scores of up to 0.72.
This corresponds to a mere 25% decrease in performance despite a nearly 1000-fold decrease in the dataset size.
- Score: 3.9146761527401424
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning has seen tremendous interest in medical imaging, particularly
in the use of convolutional neural networks (CNNs) for developing automated
diagnostic tools. The facility of its non-invasive acquisition makes retinal
fundus imaging amenable to such automated approaches. Recent work in analyzing
fundus images using CNNs relies on access to massive data for training and
validation - hundreds of thousands of images. However, data residency and data
privacy restrictions stymie the applicability of this approach in medical
settings where patient confidentiality is a mandate. Here, we showcase results
for the performance of DL on small datasets to classify patient sex from fundus
images - a trait thought not to be present or quantifiable in fundus images
until recently. We fine-tune a Resnet-152 model whose last layer has been
modified for binary classification. In several experiments, we assess
performance in the small dataset context using one private (DOVS) and one
public (ODIR) data source. Our models, developed using approximately 2500
fundus images, achieved test AUC scores of up to 0.72 (95% CI: [0.67, 0.77]).
This corresponds to a mere 25% decrease in performance despite a nearly
1000-fold decrease in the dataset size compared to prior work in the
literature. Even with a hard task like sex categorization from retinal images,
we find that classification is possible with very small datasets. Additionally,
we perform domain adaptation experiments between DOVS and ODIR; explore the
effect of data curation on training and generalizability; and investigate model
ensembling to maximize CNN classifier performance in the context of small
development datasets.
Related papers
- Local-to-Global Self-Supervised Representation Learning for Diabetic Retinopathy Grading [0.0]
This research aims to present a novel hybrid learning model using self-supervised learning and knowledge distillation.
In our algorithm, for the first time among all self-supervised learning and knowledge distillation models, the test dataset is 50% larger than the training dataset.
Compared to a similar state-of-the-art model, our results achieved higher accuracy and more effective representation spaces.
arXiv Detail & Related papers (2024-10-01T15:19:16Z) - Comparative Analysis of ImageNet Pre-Trained Deep Learning Models and
DINOv2 in Medical Imaging Classification [7.205610366609243]
In this paper, we performed a glioma grading task using three clinical modalities of brain MRI data.
We compared the performance of various pre-trained deep learning models, including those based on ImageNet and DINOv2.
Our findings indicate that in our clinical dataset, DINOv2's performance was not as strong as ImageNet-based pre-trained models.
arXiv Detail & Related papers (2024-02-12T11:49:08Z) - Additional Look into GAN-based Augmentation for Deep Learning COVID-19
Image Classification [57.1795052451257]
We study the dependence of the GAN-based augmentation performance on dataset size with a focus on small samples.
We train StyleGAN2-ADA with both sets and then, after validating the quality of generated images, we use trained GANs as one of the augmentations approaches in multi-class classification problems.
The GAN-based augmentation approach is found to be comparable with classical augmentation in the case of medium and large datasets but underperforms in the case of smaller datasets.
arXiv Detail & Related papers (2024-01-26T08:28:13Z) - Self-Supervised Pre-Training with Contrastive and Masked Autoencoder
Methods for Dealing with Small Datasets in Deep Learning for Medical Imaging [8.34398674359296]
Deep learning in medical imaging has the potential to minimize the risk of diagnostic errors, reduce radiologist workload, and accelerate diagnosis.
Training such deep learning models requires large and accurate datasets, with annotations for all training samples.
To address this challenge, deep learning models can be pre-trained on large image datasets without annotations using methods from the field of self-supervised learning.
arXiv Detail & Related papers (2023-08-12T11:31:01Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Vision-Language Modelling For Radiological Imaging and Reports In The
Low Data Regime [70.04389979779195]
This paper explores training medical vision-language models (VLMs) where the visual and language inputs are embedded into a common space.
We explore several candidate methods to improve low-data performance, including adapting generic pre-trained models to novel image and text domains.
Using text-to-image retrieval as a benchmark, we evaluate the performance of these methods with variable sized training datasets of paired chest X-rays and radiological reports.
arXiv Detail & Related papers (2023-03-30T18:20:00Z) - Retinal Image Segmentation with Small Datasets [25.095695898777656]
Many eye diseases like Diabetic Macular Edema (DME), Age-related Macular Degeneration (AMD) and Glaucoma manifest in the retina, can cause irreversible blindness or severely impair the central version.
The Optical Coherence Tomography ( OCT), a 3D scan of the retina, can be used to diagnose and monitor changes in the retinal anatomy.
Many Deep Learning (DL) methods have shared the success of developing an automated tool to monitor pathological changes in the retina.
arXiv Detail & Related papers (2023-03-09T08:32:14Z) - AMIGO: Sparse Multi-Modal Graph Transformer with Shared-Context
Processing for Representation Learning of Giga-pixel Images [53.29794593104923]
We present a novel concept of shared-context processing for whole slide histopathology images.
AMIGO uses the celluar graph within the tissue to provide a single representation for a patient.
We show that our model is strongly robust to missing information to an extent that it can achieve the same performance with as low as 20% of the data.
arXiv Detail & Related papers (2023-03-01T23:37:45Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - Classification of Microscopy Images of Breast Tissue: Region Duplication
based Self-Supervision vs. Off-the Shelf Deep Representations [0.0]
We propose a novel self-supervision pretext task to train a convolutional neural network (CNN) and extract domain specific features.
Results indicated that the best performance of 99% sensitivity was achieved for the deep features extracted using ResNet50 with concatenation of patch-level embedding.
arXiv Detail & Related papers (2022-02-12T14:12:13Z) - Classification of COVID-19 in CT Scans using Multi-Source Transfer
Learning [91.3755431537592]
We propose the use of Multi-Source Transfer Learning to improve upon traditional Transfer Learning for the classification of COVID-19 from CT scans.
With our multi-source fine-tuning approach, our models outperformed baseline models fine-tuned with ImageNet.
Our best performing model was able to achieve an accuracy of 0.893 and a Recall score of 0.897, outperforming its baseline Recall score by 9.3%.
arXiv Detail & Related papers (2020-09-22T11:53:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.