Harvard Glaucoma Detection and Progression: A Multimodal Multitask
Dataset and Generalization-Reinforced Semi-Supervised Learning
- URL: http://arxiv.org/abs/2308.13411v1
- Date: Fri, 25 Aug 2023 14:38:51 GMT
- Title: Harvard Glaucoma Detection and Progression: A Multimodal Multitask
Dataset and Generalization-Reinforced Semi-Supervised Learning
- Authors: Yan Luo, Min Shi, Yu Tian, Tobias Elze, Mengyu Wang
- Abstract summary: We develop a novel semi-supervised learning (SSL) model called pseudo supervisor to utilize unlabeled data.
Second, we release the Harvard Glaucoma Detection and Progression (Harvard-GDP) dataset.
This is the largest glaucoma detection dataset with 3D OCT imaging data and the first glaucoma progression forecasting dataset that is publicly available.
- Score: 16.465424871839627
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Glaucoma is the number one cause of irreversible blindness globally. A major
challenge for accurate glaucoma detection and progression forecasting is the
bottleneck of limited labeled patients with the state-of-the-art (SOTA) 3D
retinal imaging data of optical coherence tomography (OCT). To address the data
scarcity issue, this paper proposes two solutions. First, we develop a novel
generalization-reinforced semi-supervised learning (SSL) model called pseudo
supervisor to optimally utilize unlabeled data. Compared with SOTA models, the
proposed pseudo supervisor optimizes the policy of predicting pseudo labels
with unlabeled samples to improve empirical generalization. Our pseudo
supervisor model is evaluated with two clinical tasks consisting of glaucoma
detection and progression forecasting. The progression forecasting task is
evaluated both unimodally and multimodally. Our pseudo supervisor model
demonstrates superior performance than SOTA SSL comparison models. Moreover,
our model also achieves the best results on the publicly available LAG fundus
dataset. Second, we introduce the Harvard Glaucoma Detection and Progression
(Harvard-GDP) Dataset, a multimodal multitask dataset that includes data from
1,000 patients with OCT imaging data, as well as labels for glaucoma detection
and progression. This is the largest glaucoma detection dataset with 3D OCT
imaging data and the first glaucoma progression forecasting dataset that is
publicly available. Detailed sex and racial analysis are provided, which can be
used by interested researchers for fairness learning studies. Our released
dataset is benchmarked with several SOTA supervised CNN and transformer deep
learning models. The dataset and code are made publicly available via
\url{https://ophai.hms.harvard.edu/datasets/harvard-gdp1000}.
Related papers
- Multi-OCT-SelfNet: Integrating Self-Supervised Learning with Multi-Source Data Fusion for Enhanced Multi-Class Retinal Disease Classification [2.5091334993691206]
Development of a robust deep-learning model for retinal disease diagnosis requires a substantial dataset for training.
The capacity to generalize effectively on smaller datasets remains a persistent challenge.
We've combined a wide range of data sources to improve performance and generalization to new data.
arXiv Detail & Related papers (2024-09-17T17:22:35Z) - Automated Detection of Myopic Maculopathy in MMAC 2023: Achievements in
Classification, Segmentation, and Spherical Equivalent Prediction [6.993091116816899]
Myopic macular degeneration is the primary cause of vision loss in individuals with pathological myopia.
Early detection and prompt treatment are crucial in preventing vision impairment due to myopic maculopathy.
This was the focus of the Myopic Maculopathy Analysis Challenge (MMAC)
arXiv Detail & Related papers (2024-01-08T00:26:21Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Leveraging Semi-Supervised Graph Learning for Enhanced Diabetic
Retinopathy Detection [0.0]
Diabetic Retinopathy (DR) is a significant cause of blindness globally, highlighting the urgent need for early detection and effective treatment.
Recent advancements in Machine Learning (ML) techniques have shown promise in DR detection, but the availability of labeled data often limits their performance.
This research proposes a novel Semi-Supervised Graph Learning SSGL algorithm tailored for DR detection.
arXiv Detail & Related papers (2023-09-02T04:42:08Z) - Domain Adaptive Synapse Detection with Weak Point Annotations [63.97144211520869]
We present AdaSyn, a framework for domain adaptive synapse detection with weak point annotations.
In the WASPSYN challenge at I SBI 2023, our method ranks the 1st place.
arXiv Detail & Related papers (2023-08-31T05:05:53Z) - Augment and Criticize: Exploring Informative Samples for Semi-Supervised
Monocular 3D Object Detection [64.65563422852568]
We improve the challenging monocular 3D object detection problem with a general semi-supervised framework.
We introduce a novel, simple, yet effective Augment and Criticize' framework that explores abundant informative samples from unlabeled data.
The two new detectors, dubbed 3DSeMo_DLE and 3DSeMo_FLEX, achieve state-of-the-art results with remarkable improvements for over 3.5% AP_3D/BEV (Easy) on KITTI.
arXiv Detail & Related papers (2023-03-20T16:28:15Z) - Unified Visual Relationship Detection with Vision and Language Models [89.77838890788638]
This work focuses on training a single visual relationship detector predicting over the union of label spaces from multiple datasets.
We propose UniVRD, a novel bottom-up method for Unified Visual Relationship Detection by leveraging vision and language models.
Empirical results on both human-object interaction detection and scene-graph generation demonstrate the competitive performance of our model.
arXiv Detail & Related papers (2023-03-16T00:06:28Z) - VAESim: A probabilistic approach for self-supervised prototype discovery [0.23624125155742057]
We propose an architecture for image stratification based on a conditional variational autoencoder.
We use a continuous latent space to represent the continuum of disorders and find clusters during training, which can then be used for image/patient stratification.
We demonstrate that our method outperforms baselines in terms of kNN accuracy measured on a classification task against a standard VAE.
arXiv Detail & Related papers (2022-09-25T17:55:31Z) - A Real Use Case of Semi-Supervised Learning for Mammogram Classification
in a Local Clinic of Costa Rica [0.5541644538483946]
Training a deep learning model requires a considerable amount of labeled images.
A number of publicly available datasets have been built with data from different hospitals and clinics.
The use of the semi-supervised deep learning approach known as MixMatch, to leverage the usage of unlabeled data is proposed and evaluated.
arXiv Detail & Related papers (2021-07-24T22:26:50Z) - Modeling Shared Responses in Neuroimaging Studies through MultiView ICA [94.31804763196116]
Group studies involving large cohorts of subjects are important to draw general conclusions about brain functional organization.
We propose a novel MultiView Independent Component Analysis model for group studies, where data from each subject are modeled as a linear combination of shared independent sources plus noise.
We demonstrate the usefulness of our approach first on fMRI data, where our model demonstrates improved sensitivity in identifying common sources among subjects.
arXiv Detail & Related papers (2020-06-11T17:29:53Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.