Extrapolating Prospective Glaucoma Fundus Images through Diffusion Model in Irregular Longitudinal Sequences
- URL: http://arxiv.org/abs/2410.21130v2
- Date: Tue, 05 Nov 2024 10:22:35 GMT
- Title: Extrapolating Prospective Glaucoma Fundus Images through Diffusion Model in Irregular Longitudinal Sequences
- Authors: Zhihao Zhao, Junjie Yang, Shahrooz Faghihroohi, Yinzheng Zhao, Daniel Zapp, Kai Huang, Nassir Navab, M. Ali Nasseri,
- Abstract summary: The utilization of longitudinal datasets for glaucoma progression prediction offers a compelling approach to support early therapeutic interventions.
We propose a novel diffusion-based model to predict prospective images by extrapolating from existing longitudinal fundus images of patients.
- Score: 46.80977922491862
- License:
- Abstract: The utilization of longitudinal datasets for glaucoma progression prediction offers a compelling approach to support early therapeutic interventions. Predominant methodologies in this domain have primarily focused on the direct prediction of glaucoma stage labels from longitudinal datasets. However, such methods may not adequately encapsulate the nuanced developmental trajectory of the disease. To enhance the diagnostic acumen of medical practitioners, we propose a novel diffusion-based model to predict prospective images by extrapolating from existing longitudinal fundus images of patients. The methodology delineated in this study distinctively leverages sequences of images as inputs. Subsequently, a time-aligned mask is employed to select a specific year for image generation. During the training phase, the time-aligned mask resolves the issue of irregular temporal intervals in longitudinal image sequence sampling. Additionally, we utilize a strategy of randomly masking a frame in the sequence to establish the ground truth. This methodology aids the network in continuously acquiring knowledge regarding the internal relationships among the sequences throughout the learning phase. Moreover, the introduction of textual labels is instrumental in categorizing images generated within the sequence. The empirical findings from the conducted experiments indicate that our proposed model not only effectively generates longitudinal data but also significantly improves the precision of downstream classification tasks.
Related papers
- Spatial-aware Attention Generative Adversarial Network for Semi-supervised Anomaly Detection in Medical Image [63.59114880750643]
We introduce a novel Spatial-aware Attention Generative Adrialversa Network (SAGAN) for one-class semi-supervised generation of health images.
SAGAN generates high-quality health images corresponding to unlabeled data, guided by the reconstruction of normal images and restoration of pseudo-anomaly images.
Extensive experiments on three medical datasets demonstrate that the proposed SAGAN outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2024-05-21T15:41:34Z) - Harnessing the power of longitudinal medical imaging for eye disease prognosis using Transformer-based sequence modeling [49.52787013516891]
Our proposed Longitudinal Transformer for Survival Analysis (LTSA) enables dynamic disease prognosis from longitudinal medical imaging.
A temporal attention analysis also suggested that, while the most recent image is typically the most influential, prior imaging still provides additional prognostic value.
arXiv Detail & Related papers (2024-05-14T17:15:28Z) - Metadata-enhanced contrastive learning from retinal optical coherence tomography images [7.932410831191909]
We extend conventional contrastive frameworks with a novel metadata-enhanced strategy.
Our approach employs widely available patient metadata to approximate the true set of inter-image contrastive relationships.
Our approach outperforms both standard contrastive methods and a retinal image foundation model in five out of six image-level downstream tasks.
arXiv Detail & Related papers (2022-08-04T08:53:15Z) - Local Spatiotemporal Representation Learning for
Longitudinally-consistent Neuroimage Analysis [7.568469725821069]
This paper presents a local and multi-scaletemporal representation learning method for image-to-image architectures trained on longitudinal images.
During finetuning, it proposes a surprisingly simple self-supervised segmentation consistency regularization to exploit intrasubject correlation.
These improvements are demonstrated across both longitudinal neurodegenerative adult and developing infant brain MRI and yield both higher performance and longitudinal consistency.
arXiv Detail & Related papers (2022-06-09T05:17:00Z) - Temporal Context Matters: Enhancing Single Image Prediction with Disease
Progression Representations [8.396615243014768]
We present a deep learning approach that leverages temporal progression information to improve clinical outcome predictions from single-timepoint images.
In our method, a self-attention based Temporal Convolutional Network (TCN) is used to learn a representation that is most reflective of the disease trajectory.
A Vision Transformer is pretrained in a self-supervised fashion to extract features from single-timepoint images.
arXiv Detail & Related papers (2022-03-02T22:11:07Z) - Learning Discriminative Shrinkage Deep Networks for Image Deconvolution [122.79108159874426]
We propose an effective non-blind deconvolution approach by learning discriminative shrinkage functions to implicitly model these terms.
Experimental results show that the proposed method performs favorably against the state-of-the-art ones in terms of efficiency and accuracy.
arXiv Detail & Related papers (2021-11-27T12:12:57Z) - Pseudo-domains in imaging data improve prediction of future disease
status in multi-center studies [57.712855968194305]
We propose a prediction method that can cope with a high number of different scanning sites and a low number of samples per site.
Results show that they improve the prediction accuracy for steatosis after 48 weeks from imaging data acquired at an initial visit and 12-weeks follow-up in liver disease.
arXiv Detail & Related papers (2021-11-15T09:40:54Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Development and Validation of a Novel Prognostic Model for Predicting
AMD Progression Using Longitudinal Fundus Images [6.258161719849178]
We propose a novel deep learning method to predict the progression of diseases using longitudinal imaging data with uneven time intervals.
We demonstrate our method on a longitudinal dataset of color fundus images from 4903 eyes with age-related macular degeneration (AMD)
Our method attains a testing sensitivity of 0.878, a specificity of 0.887, and an area under the receiver operating characteristic of 0.950.
arXiv Detail & Related papers (2020-07-10T00:33:19Z) - Unsupervised Lesion Detection via Image Restoration with a Normative
Prior [6.495883501989547]
We propose a probabilistic model that uses a network-based prior as the normative distribution and detect lesions pixel-wise using MAP estimation.
Experiments with gliomas and stroke lesions in brain MRI show that the proposed approach outperforms the state-of-the-art unsupervised methods by a substantial margin.
arXiv Detail & Related papers (2020-04-30T18:03:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.