Camera Adaptation for Fundus-Image-Based CVD Risk Estimation
- URL: http://arxiv.org/abs/2206.09202v1
- Date: Sat, 18 Jun 2022 13:28:16 GMT
- Title: Camera Adaptation for Fundus-Image-Based CVD Risk Estimation
- Authors: Zhihong Lin, Danli Shi, Donghao Zhang, Xianwen Shang, Mingguang He,
Zongyuan Ge
- Abstract summary: Combining deep learning (DL) and portable fundus cameras will enable CVD risk estimation in various scenarios.
One of the top priority issues is the different camera differences between the databases for research material and the samples in the production environment.
We propose a cross-laterality feature alignment pre-training scheme and a self-attention camera adaptor module to improve the model robustness.
- Score: 20.240895185459618
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent studies have validated the association between cardiovascular disease
(CVD) risk and retinal fundus images. Combining deep learning (DL) and portable
fundus cameras will enable CVD risk estimation in various scenarios and improve
healthcare democratization. However, there are still significant issues to be
solved. One of the top priority issues is the different camera differences
between the databases for research material and the samples in the production
environment. Most high-quality retinography databases ready for research are
collected from high-end fundus cameras, and there is a significant domain
discrepancy between different cameras. To fully explore the domain discrepancy
issue, we first collect a Fundus Camera Paired (FCP) dataset containing
pair-wise fundus images captured by the high-end Topcon retinal camera and the
low-end Mediwork portable fundus camera of the same patients. Then, we propose
a cross-laterality feature alignment pre-training scheme and a self-attention
camera adaptor module to improve the model robustness. The cross-laterality
feature alignment training encourages the model to learn common knowledge from
the same patient's left and right fundus images and improve model
generalization. Meanwhile, the device adaptation module learns feature
transformation from the target domain to the source domain. We conduct
comprehensive experiments on both the UK Biobank database and our FCP data. The
experimental results show that the CVD risk regression accuracy and the result
consistency over two cameras are improved with our proposed method. The code is
available here:
\url{https://github.com/linzhlalala/CVD-risk-based-on-retinal-fundus-images}
Related papers
- Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - RIDE: Self-Supervised Learning of Rotation-Equivariant Keypoint
Detection and Invariant Description for Endoscopy [83.4885991036141]
RIDE is a learning-based method for rotation-equivariant detection and invariant description.
It is trained in a self-supervised manner on a large curation of endoscopic images.
It sets a new state-of-the-art performance on matching and relative pose estimation tasks.
arXiv Detail & Related papers (2023-09-18T08:16:30Z) - Cross-Site Severity Assessment of COVID-19 from CT Images via Domain
Adaptation [64.59521853145368]
Early and accurate severity assessment of Coronavirus disease 2019 (COVID-19) based on computed tomography (CT) images offers a great help to the estimation of intensive care unit event.
To augment the labeled data and improve the generalization ability of the classification model, it is necessary to aggregate data from multiple sites.
This task faces several challenges including class imbalance between mild and severe infections, domain distribution discrepancy between sites, and presence of heterogeneous features.
arXiv Detail & Related papers (2021-09-08T07:56:51Z) - DRDrV3: Complete Lesion Detection in Fundus Images Using Mask R-CNN,
Transfer Learning, and LSTM [2.9360071145551068]
We propose a new lesion detection architecture, comprising of two sub-modules, which is an optimal solution to detect and find lesions caused by Diabetic Retinopathy (DR)
We also use two popular evaluation criteria to evaluate the outputs of our models, which are intersection over union (IOU) and mean average precision (mAP)
We hypothesize that this new solution enables specialists to detect lesions with high confidence and estimate the severity of the damage with high accuracy.
arXiv Detail & Related papers (2021-08-18T11:36:37Z) - Multi-frame Collaboration for Effective Endoscopic Video Polyp Detection
via Spatial-Temporal Feature Transformation [28.01363432141765]
We present Spatial-Temporal Feature Transformation (STFT), a multi-frame collaborative framework to address these issues.
For example, STFT mitigates inter-frame variations in the camera-moving situation with feature alignment by proposal-guided deformable convolutions.
Empirical studies and superior results demonstrate the effectiveness and stability of our method.
arXiv Detail & Related papers (2021-07-08T05:17:30Z) - Self-Adaptive Transfer Learning for Multicenter Glaucoma Classification
in Fundus Retina Images [9.826586293806837]
We propose a self-adaptive transfer learning (SATL) strategy to fill the domain gap between multicenter datasets.
Specifically, the encoder of a DL model that is pre-trained on the source domain is used to initialize the encoder of a reconstruction model.
Results demonstrate that the proposed SATL strategy is effective in the domain adaptation task between a private and two public glaucoma diagnosis datasets.
arXiv Detail & Related papers (2021-05-07T05:20:37Z) - Joint Noise-Tolerant Learning and Meta Camera Shift Adaptation for
Unsupervised Person Re-Identification [60.36551512902312]
unsupervised person re-identification (re-ID) aims to learn discriminative models with unlabeled data.
One popular method is to obtain pseudo-label by clustering and use them to optimize the model.
In this paper, we propose a unified framework to solve both problems.
arXiv Detail & Related papers (2021-03-08T09:13:06Z) - Fader Networks for domain adaptation on fMRI: ABIDE-II study [68.5481471934606]
We use 3D convolutional autoencoders to build the domain irrelevant latent space image representation and demonstrate this method to outperform existing approaches on ABIDE data.
arXiv Detail & Related papers (2020-10-14T16:50:50Z) - Robust Retinal Vessel Segmentation from a Data Augmentation Perspective [14.768009562830004]
We propose two new data augmentation modules, namely, channel-wise random Gamma correction and channel-wise random vessel augmentation.
With the additional training samples generated by applying these two modules sequentially, a model could learn more invariant and discriminating features.
Experimental results on both real-world and synthetic datasets demonstrate that our method can improve the performance and robustness of a classic convolutional neural network architecture.
arXiv Detail & Related papers (2020-07-31T07:37:14Z) - Improved Slice-wise Tumour Detection in Brain MRIs by Computing
Dissimilarities between Latent Representations [68.8204255655161]
Anomaly detection for Magnetic Resonance Images (MRIs) can be solved with unsupervised methods.
We have proposed a slice-wise semi-supervised method for tumour detection based on the computation of a dissimilarity function in the latent space of a Variational AutoEncoder.
We show that by training the models on higher resolution images and by improving the quality of the reconstructions, we obtain results which are comparable with different baselines.
arXiv Detail & Related papers (2020-07-24T14:02:09Z) - Target-Independent Domain Adaptation for WBC Classification using
Generative Latent Search [20.199195698983715]
Unsupervised Domain Adaptation (UDA) techniques presuppose the existence of sufficient amount of unlabelled target data.
We propose a method for UDA that is devoid of the need for target data.
We prove the existence of such a clone given that infinite number of data points can be sampled from the source distribution.
arXiv Detail & Related papers (2020-05-11T20:58:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.