Supervised Domain Adaptation for Recognizing Retinal Diseases from
Wide-Field Fundus Images
- URL: http://arxiv.org/abs/2305.08078v2
- Date: Tue, 24 Oct 2023 02:12:02 GMT
- Title: Supervised Domain Adaptation for Recognizing Retinal Diseases from
Wide-Field Fundus Images
- Authors: Qijie Wei, Jingyuan Yang, Bo Wang, Jinrui Wang, Jianchun Zhao, Xinyu
Zhao, Sheng Yang, Niranchana Manivannan, Youxin Chen, Dayong Ding, Jing Zhou
and Xirong Li
- Abstract summary: This paper addresses the emerging task of recognizing multiple retinal diseases from wide-field (WF) and ultra-wide-field (UWF) fundus images.
We propose a supervised domain adaptation method named Cross-domain Collaborative Learning (CdCL)
Inspired by the success of fixed-based mixup in unsupervised domain adaptation, we re-purpose this strategy for the current task.
- Score: 23.503104144297684
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper addresses the emerging task of recognizing multiple retinal
diseases from wide-field (WF) and ultra-wide-field (UWF) fundus images. For an
effective use of existing large amount of labeled color fundus photo (CFP) data
and the relatively small amount of WF and UWF data, we propose a supervised
domain adaptation method named Cross-domain Collaborative Learning (CdCL).
Inspired by the success of fixed-ratio based mixup in unsupervised domain
adaptation, we re-purpose this strategy for the current task. Due to the
intrinsic disparity between the field-of-view of CFP and WF/UWF images, a scale
bias naturally exists in a mixup sample that the anatomic structure from a CFP
image will be considerably larger than its WF/UWF counterpart. The CdCL method
resolves the issue by Scale-bias Correction, which employs Transformers for
producing scale-invariant features. As demonstrated by extensive experiments on
multiple datasets covering both WF and UWF images, the proposed method compares
favorably against a number of competitive baselines.
Related papers
- Unsupervised Modality Adaptation with Text-to-Image Diffusion Models for Semantic Segmentation [54.96563068182733]
We propose Modality Adaptation with text-to-image Diffusion Models (MADM) for semantic segmentation task.
MADM utilizes text-to-image diffusion models pre-trained on extensive image-text pairs to enhance the model's cross-modality capabilities.
We show that MADM achieves state-of-the-art adaptation performance across various modality tasks, including images to depth, infrared, and event modalities.
arXiv Detail & Related papers (2024-10-29T03:49:40Z) - UWAFA-GAN: Ultra-Wide-Angle Fluorescein Angiography Transformation via Multi-scale Generation and Registration Enhancement [17.28459176559761]
UWF fluorescein angiography (UWF-FA) requires the administration of a fluorescent dye via injection into the patient's hand or elbow.
To mitigate potential adverse effects associated with injections, researchers have proposed the development of cross-modality medical image generation algorithms.
We introduce a novel conditional generative adversarial network (UWAFA-GAN) to synthesize UWF-FA from UWF-SLO.
arXiv Detail & Related papers (2024-05-01T14:27:43Z) - Misalignment-Robust Frequency Distribution Loss for Image Transformation [51.0462138717502]
This paper aims to address a common challenge in deep learning-based image transformation methods, such as image enhancement and super-resolution.
We introduce a novel and simple Frequency Distribution Loss (FDL) for computing distribution distance within the frequency domain.
Our method is empirically proven effective as a training constraint due to the thoughtful utilization of global information in the frequency domain.
arXiv Detail & Related papers (2024-02-28T09:27:41Z) - Improving Misaligned Multi-modality Image Fusion with One-stage
Progressive Dense Registration [67.23451452670282]
Misalignments between multi-modality images pose challenges in image fusion.
We propose a Cross-modality Multi-scale Progressive Dense Registration scheme.
This scheme accomplishes the coarse-to-fine registration exclusively using a one-stage optimization.
arXiv Detail & Related papers (2023-08-22T03:46:24Z) - UWAT-GAN: Fundus Fluorescein Angiography Synthesis via Ultra-wide-angle
Transformation Multi-scale GAN [1.165405976310311]
Fundus photography is an essential examination for clinical and differential diagnosis of fundus diseases.
Current methods in fundus imaging could not produce high-resolution images and are unable to capture tiny vascular lesion areas.
This paper proposes a novel conditional generative adversarial network (UWAT-GAN) to synthesize UWF-FA from UWF-SLO.
arXiv Detail & Related papers (2023-07-21T12:23:39Z) - MSCDA: Multi-level Semantic-guided Contrast Improves Unsupervised Domain
Adaptation for Breast MRI Segmentation in Small Datasets [5.272836235045653]
We propose a novel Multi-level Semantic-guided Contrastive Domain Adaptation framework.
Our approach incorporates self-training with contrastive learning to align feature representations between domains.
In particular, we extend the contrastive loss by incorporating pixel-to-pixel, pixel-to-centroid, and centroid-to-centroid contrasts.
arXiv Detail & Related papers (2023-01-04T19:16:55Z) - Multi-Scale Multi-Target Domain Adaptation for Angle Closure
Classification [50.658613573816254]
We propose a novel Multi-scale Multi-target Domain Adversarial Network (M2DAN) for angle closure classification.
Based on these domain-invariant features at different scales, the deep model trained on the source domain is able to classify angle closure on multiple target domains.
arXiv Detail & Related papers (2022-08-25T15:27:55Z) - PIT: Position-Invariant Transform for Cross-FoV Domain Adaptation [53.428312630479816]
We observe that the Field of View (FoV) gap induces noticeable instance appearance differences between the source and target domains.
Motivated by the observations, we propose the textbfPosition-Invariant Transform (PIT) to better align images in different domains.
arXiv Detail & Related papers (2021-08-16T15:16:47Z) - Leveraging Regular Fundus Images for Training UWF Fundus Diagnosis
Models via Adversarial Learning and Pseudo-Labeling [29.009663623719064]
Ultra-widefield (UWF) 200degreefundus imaging by Optos cameras has gradually been introduced.
Regular fundus images contain a large amount of high-quality and well-annotated data.
Due to the domain gap, models trained by regular fundus images to recognize UWF fundus images perform poorly.
We propose the use of a modified cycle generative adversarial network (CycleGAN) model to bridge the gap between regular and UWF fundus.
arXiv Detail & Related papers (2020-11-27T16:25:30Z) - Deep Semantic Matching with Foreground Detection and Cycle-Consistency [103.22976097225457]
We address weakly supervised semantic matching based on a deep network.
We explicitly estimate the foreground regions to suppress the effect of background clutter.
We develop cycle-consistent losses to enforce the predicted transformations across multiple images to be geometrically plausible and consistent.
arXiv Detail & Related papers (2020-03-31T22:38:09Z) - Bridge the Domain Gap Between Ultra-wide-field and Traditional Fundus
Images via Adversarial Domain Adaptation [38.697737290155715]
Ultra-wide-field fundus imaging by Optos camera is gradually put into use because of its broader insights on fundus.
Research on traditional fundus images is an active topic but studies on UWF fundus images are few.
We propose a flexible framework to bridge the domain gap between two domains and co-train a UWF fundus diagnosis model by pseudo-labelling and adversarial learning.
arXiv Detail & Related papers (2020-03-23T01:31:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.