Exploring the Transferability of a Foundation Model for Fundus Images:
Application to Hypertensive Retinopathy
- URL: http://arxiv.org/abs/2401.15526v1
- Date: Sat, 27 Jan 2024 23:40:24 GMT
- Title: Exploring the Transferability of a Foundation Model for Fundus Images:
Application to Hypertensive Retinopathy
- Authors: Julio Silva-Rodriguez, Jihed Chelbi, Waziha Kabir, Hadi Chakor, Jose
Dolz, Ismail Ben Ayed and Riadh Kobbi
- Abstract summary: Using deep learning models pre-trained on Imagenet is the traditional solution for medical image classification to deal with data scarcity.
The CGI-HRDC challenge for Hypertensive Retinopathy diagnosis on fundus images introduces an appealing opportunity to evaluate the transferability of a recently released vision-language foundation model of the retina, FLAIR.
- Score: 15.643435527710817
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Using deep learning models pre-trained on Imagenet is the traditional
solution for medical image classification to deal with data scarcity.
Nevertheless, relevant literature supports that this strategy may offer limited
gains due to the high dissimilarity between domains. Currently, the paradigm of
adapting domain-specialized foundation models is proving to be a promising
alternative. However, how to perform such knowledge transfer, and the benefits
and limitations it presents, are under study. The CGI-HRDC challenge for
Hypertensive Retinopathy diagnosis on fundus images introduces an appealing
opportunity to evaluate the transferability of a recently released
vision-language foundation model of the retina, FLAIR. In this work, we explore
the potential of using FLAIR features as starting point for fundus image
classification, and we compare its performance with regard to Imagenet
initialization on two popular transfer learning methods: Linear Probing (LP)
and Fine-Tuning (FP). Our empirical observations suggest that, in any case, the
use of the traditional strategy provides performance gains. In contrast, direct
transferability from FLAIR model allows gains of 2.5%. When fine-tuning the
whole network, the performance gap increases up to 4%. In this case, we show
that avoiding feature deterioration via LP initialization of the classifier
allows the best re-use of the rich pre-trained features. Although direct
transferability using LP still offers limited performance, we believe that
foundation models such as FLAIR will drive the evolution of deep-learning-based
fundus image analysis.
Related papers
- JoReS-Diff: Joint Retinex and Semantic Priors in Diffusion Model for Low-light Image Enhancement [69.6035373784027]
Low-light image enhancement (LLIE) has achieved promising performance by employing conditional diffusion models.
Previous methods may neglect the importance of a sufficient formulation of task-specific condition strategy.
We propose JoReS-Diff, a novel approach that incorporates Retinex- and semantic-based priors as the additional pre-processing condition.
arXiv Detail & Related papers (2023-12-20T08:05:57Z) - Forward-Forward Contrastive Learning [4.465144120325802]
We propose Forward Forward Contrastive Learning (FFCL) as a novel pretraining approach for medical image classification.
FFCL achieves superior performance (3.69% accuracy over ImageNet pretrained ResNet-18) over existing pretraining models in the pneumonia classification task.
arXiv Detail & Related papers (2023-05-04T15:29:06Z) - Performance of GAN-based augmentation for deep learning COVID-19 image
classification [57.1795052451257]
The biggest challenge in the application of deep learning to the medical domain is the availability of training data.
Data augmentation is a typical methodology used in machine learning when confronted with a limited data set.
In this work, a StyleGAN2-ADA model of Generative Adversarial Networks is trained on the limited COVID-19 chest X-ray image set.
arXiv Detail & Related papers (2023-04-18T15:39:58Z) - Towards Foundation Models and Few-Shot Parameter-Efficient Fine-Tuning for Volumetric Organ Segmentation [20.94974284175104]
Few-Shot Efficient Fine-Tuning (FSEFT) is a novel and realistic scenario for adapting medical image segmentation foundation models.
Our comprehensive transfer learning experiments confirm the suitability of foundation models in medical image segmentation and unveil the limitations of popular fine-tuning strategies in few-shot scenarios.
arXiv Detail & Related papers (2023-03-29T22:50:05Z) - Unsupervised Domain Transfer with Conditional Invertible Neural Networks [83.90291882730925]
We propose a domain transfer approach based on conditional invertible neural networks (cINNs)
Our method inherently guarantees cycle consistency through its invertible architecture, and network training can efficiently be conducted with maximum likelihood.
Our method enables the generation of realistic spectral data and outperforms the state of the art on two downstream classification tasks.
arXiv Detail & Related papers (2023-03-17T18:00:27Z) - Bridging Synthetic and Real Images: a Transferable and Multiple
Consistency aided Fundus Image Enhancement Framework [61.74188977009786]
We propose an end-to-end optimized teacher-student framework to simultaneously conduct image enhancement and domain adaptation.
We also propose a novel multi-stage multi-attention guided enhancement network (MAGE-Net) as the backbones of our teacher and student network.
arXiv Detail & Related papers (2023-02-23T06:16:15Z) - Pre-text Representation Transfer for Deep Learning with Limited
Imbalanced Data : Application to CT-based COVID-19 Detection [18.72489078928417]
We propose a novel concept of Pre-text Representation Transfer (PRT)
PRT retains the original classification layers and updates the representation layers through an unsupervised pre-text task.
Our results show a consistent gain over the conventional transfer learning with the proposed method.
arXiv Detail & Related papers (2023-01-21T04:47:35Z) - Learning Discriminative Shrinkage Deep Networks for Image Deconvolution [122.79108159874426]
We propose an effective non-blind deconvolution approach by learning discriminative shrinkage functions to implicitly model these terms.
Experimental results show that the proposed method performs favorably against the state-of-the-art ones in terms of efficiency and accuracy.
arXiv Detail & Related papers (2021-11-27T12:12:57Z) - Differentially private federated deep learning for multi-site medical
image segmentation [56.30543374146002]
Collaborative machine learning techniques such as federated learning (FL) enable the training of models on effectively larger datasets without data transfer.
Recent initiatives have demonstrated that segmentation models trained with FL can achieve performance similar to locally trained models.
However, FL is not a fully privacy-preserving technique and privacy-centred attacks can disclose confidential patient data.
arXiv Detail & Related papers (2021-07-06T12:57:32Z) - Leveraging Regular Fundus Images for Training UWF Fundus Diagnosis
Models via Adversarial Learning and Pseudo-Labeling [29.009663623719064]
Ultra-widefield (UWF) 200degreefundus imaging by Optos cameras has gradually been introduced.
Regular fundus images contain a large amount of high-quality and well-annotated data.
Due to the domain gap, models trained by regular fundus images to recognize UWF fundus images perform poorly.
We propose the use of a modified cycle generative adversarial network (CycleGAN) model to bridge the gap between regular and UWF fundus.
arXiv Detail & Related papers (2020-11-27T16:25:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.