COROLLA: An Efficient Multi-Modality Fusion Framework with Supervised
Contrastive Learning for Glaucoma Grading
- URL: http://arxiv.org/abs/2201.03795v1
- Date: Tue, 11 Jan 2022 06:00:51 GMT
- Title: COROLLA: An Efficient Multi-Modality Fusion Framework with Supervised
Contrastive Learning for Glaucoma Grading
- Authors: Zhiyuan Cai, Li Lin, Huaqing He, Xiaoying Tang
- Abstract summary: We propose an efficient multi-modality supervised contrastive learning framework, named COROLLA, for glaucoma grading.
We employ supervised contrastive learning to increase our models' discriminative power with better convergence.
On the GAMMA dataset, our COROLLA framework achieves overwhelming glaucoma grading performance compared to state-of-the-art methods.
- Score: 1.2250035750661867
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Glaucoma is one of the ophthalmic diseases that may cause blindness, for
which early detection and treatment are very important. Fundus images and
optical coherence tomography (OCT) images are both widely-used modalities in
diagnosing glaucoma. However, existing glaucoma grading approaches mainly
utilize a single modality, ignoring the complementary information between
fundus and OCT. In this paper, we propose an efficient multi-modality
supervised contrastive learning framework, named COROLLA, for glaucoma grading.
Through layer segmentation as well as thickness calculation and projection,
retinal thickness maps are extracted from the original OCT volumes and used as
a replacing modality, resulting in more efficient calculations with less memory
usage. Given the high structure and distribution similarities across medical
image samples, we employ supervised contrastive learning to increase our
models' discriminative power with better convergence. Moreover, feature-level
fusion of paired fundus image and thickness map is conducted for enhanced
diagnosis accuracy. On the GAMMA dataset, our COROLLA framework achieves
overwhelming glaucoma grading performance compared to state-of-the-art methods.
Related papers
- ETSCL: An Evidence Theory-Based Supervised Contrastive Learning Framework for Multi-modal Glaucoma Grading [7.188153974946432]
Glaucoma is one of the leading causes of vision impairment.
It remains challenging to extract reliable features due to the high similarity of medical images and the unbalanced multi-modal data distribution.
We propose a novel framework, namely ETSCL, which consists of a contrastive feature extraction stage and a decision-level fusion stage.
arXiv Detail & Related papers (2024-07-19T11:57:56Z) - A better approach to diagnose retinal diseases: Combining our Segmentation-based Vascular Enhancement with deep learning features [3.717366858126521]
Abnormalities in retinal fundus images may indicate certain pathologies.
In conventional medicine, the diagnosis of retina-related diseases relies on a physician's subjective assessment of the retinal fundus images.
This paper proposes a fast, objective, and accurate method for the diagnosis of diseases related to retinal fundus images.
arXiv Detail & Related papers (2024-05-25T13:52:40Z) - ELF: An End-to-end Local and Global Multimodal Fusion Framework for
Glaucoma Grading [43.12236694270165]
We propose an end-to-end local and global multi-modal fusion framework for glaucoma grading named ELF.
ELF can fully utilize the complementary information between fundus and OCT.
The extensive experiment conducted on the multi-modal glaucoma grading GAMMA dataset can prove the effiectness of ELF.
arXiv Detail & Related papers (2023-11-14T09:51:00Z) - Fundus-Enhanced Disease-Aware Distillation Model for Retinal Disease
Classification from OCT Images [6.72159216082989]
We propose a fundus-enhanced disease-aware distillation model for retinal disease classification from OCT images.
Our framework enhances the OCT model during training by utilizing unpaired fundus images.
Our proposed approach outperforms single-modal, multi-modal, and state-of-the-art distillation methods for retinal disease classification.
arXiv Detail & Related papers (2023-08-01T05:13:02Z) - Bridging Synthetic and Real Images: a Transferable and Multiple
Consistency aided Fundus Image Enhancement Framework [61.74188977009786]
We propose an end-to-end optimized teacher-student framework to simultaneously conduct image enhancement and domain adaptation.
We also propose a novel multi-stage multi-attention guided enhancement network (MAGE-Net) as the backbones of our teacher and student network.
arXiv Detail & Related papers (2023-02-23T06:16:15Z) - Affinity Feature Strengthening for Accurate, Complete and Robust Vessel
Segmentation [48.638327652506284]
Vessel segmentation is crucial in many medical image applications, such as detecting coronary stenoses, retinal vessel diseases and brain aneurysms.
We present a novel approach, the affinity feature strengthening network (AFN), which jointly models geometry and refines pixel-wise segmentation features using a contrast-insensitive, multiscale affinity approach.
arXiv Detail & Related papers (2022-11-12T05:39:17Z) - GAMMA Challenge:Glaucoma grAding from Multi-Modality imAges [48.98620387924817]
We set up the Glaucoma grAding from Multi-Modality imAges (GAMMA) Challenge to encourage the development of fundus & OCT-based glaucoma grading.
The primary task of the challenge is to grade glaucoma from both the 2D fundus images and 3D OCT scanning volumes.
We have publicly released a glaucoma annotated dataset with both 2D fundus color photography and 3D OCT volumes, which is the first multi-modality dataset for glaucoma grading.
arXiv Detail & Related papers (2022-02-14T06:54:15Z) - Assessing glaucoma in retinal fundus photographs using Deep Feature
Consistent Variational Autoencoders [63.391402501241195]
glaucoma is challenging to detect since it remains asymptomatic until the symptoms are severe.
Early identification of glaucoma is generally made based on functional, structural, and clinical assessments.
Deep learning methods have partially solved this dilemma by bypassing the marker identification stage and analyzing high-level information directly to classify the data.
arXiv Detail & Related papers (2021-10-04T16:06:49Z) - Modeling and Enhancing Low-quality Retinal Fundus Images [167.02325845822276]
Low-quality fundus images increase uncertainty in clinical observation and lead to the risk of misdiagnosis.
We propose a clinically oriented fundus enhancement network (cofe-Net) to suppress global degradation factors.
Experiments on both synthetic and real images demonstrate that our algorithm effectively corrects low-quality fundus images without losing retinal details.
arXiv Detail & Related papers (2020-05-12T08:01:16Z) - Pathological Retinal Region Segmentation From OCT Images Using Geometric
Relation Based Augmentation [84.7571086566595]
We propose improvements over previous GAN-based medical image synthesis methods by jointly encoding the intrinsic relationship of geometry and shape.
The proposed method outperforms state-of-the-art segmentation methods on the public RETOUCH dataset having images captured from different acquisition procedures.
arXiv Detail & Related papers (2020-03-31T11:50:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.