Dueling Deep Q-Network for Unsupervised Inter-frame Eye Movement
Correction in Optical Coherence Tomography Volumes
- URL: http://arxiv.org/abs/2007.01522v1
- Date: Fri, 3 Jul 2020 07:14:30 GMT
- Title: Dueling Deep Q-Network for Unsupervised Inter-frame Eye Movement
Correction in Optical Coherence Tomography Volumes
- Authors: Yasmeen M. George, Suman Sedai, Bhavna J. Antony, Hiroshi Ishikawa,
Gadi Wollstein, Joel S. Schuman and Rahil Garnavi
- Abstract summary: In optical coherence tomography ( OCT) volumes of retina, the sequential acquisition of the individual slices makes this modality prone to motion artifacts.
Speckle noise that is characteristic of this imaging modality, leads to inaccuracies when traditional registration techniques are employed.
In this paper, we tackle these issues by using deep reinforcement learning to correct inter-frame movements in an unsupervised manner.
- Score: 5.371290280449071
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In optical coherence tomography (OCT) volumes of retina, the sequential
acquisition of the individual slices makes this modality prone to motion
artifacts, misalignments between adjacent slices being the most noticeable. Any
distortion in OCT volumes can bias structural analysis and influence the
outcome of longitudinal studies. On the other hand, presence of speckle noise
that is characteristic of this imaging modality, leads to inaccuracies when
traditional registration techniques are employed. Also, the lack of a
well-defined ground truth makes supervised deep-learning techniques ill-posed
to tackle the problem. In this paper, we tackle these issues by using deep
reinforcement learning to correct inter-frame movements in an unsupervised
manner. Specifically, we use dueling deep Q-network to train an artificial
agent to find the optimal policy, i.e. a sequence of actions, that best
improves the alignment by maximizing the sum of reward signals. Instead of
relying on the ground-truth of transformation parameters to guide the rewarding
system, for the first time, we use a combination of intensity based image
similarity metrics. Further, to avoid the agent bias towards speckle noise, we
ensure the agent can see retinal layers as part of the interacting environment.
For quantitative evaluation, we simulate the eye movement artifacts by applying
2D rigid transformations on individual B-scans. The proposed model achieves an
average of 0.985 and 0.914 for normalized mutual information and correlation
coefficient, respectively. We also compare our model with elastix intensity
based medical image registration approach, where significant improvement is
achieved by our model for both noisy and denoised volumes.
Related papers
- CeViT: Copula-Enhanced Vision Transformer in multi-task learning and bi-group image covariates with an application to myopia screening [9.928208927136874]
We present a Vision Transformer-based bi-channel architecture, named CeViT, where the common features of a pair of eyes are extracted via a shared Transformer encoder.
We demonstrate that CeViT enhances the baseline model in both accuracy of classifying high-myopia and prediction of AL on both eyes.
arXiv Detail & Related papers (2025-01-11T13:23:56Z) - Understanding and Improving Training-Free AI-Generated Image Detections with Vision Foundation Models [68.90917438865078]
Deepfake techniques for facial synthesis and editing pose serious risks for generative models.
In this paper, we investigate how detection performance varies across model backbones, types, and datasets.
We introduce Contrastive Blur, which enhances performance on facial images, and MINDER, which addresses noise type bias, balancing performance across domains.
arXiv Detail & Related papers (2024-11-28T13:04:45Z) - Deformation-aware GAN for Medical Image Synthesis with Substantially Misaligned Pairs [0.0]
We propose a novel Deformation-aware GAN (DA-GAN) to dynamically correct the misalignment during the image synthesis based on inverse consistency.
Experimental results show that DA-GAN achieved superior performance on a public dataset with simulated misalignments and a real-world lung MRI-CT dataset with respiratory motion misalignment.
arXiv Detail & Related papers (2024-08-18T10:29:35Z) - Gadolinium dose reduction for brain MRI using conditional deep learning [66.99830668082234]
Two main challenges for these approaches are the accurate prediction of contrast enhancement and the synthesis of realistic images.
We address both challenges by utilizing the contrast signal encoded in the subtraction images of pre-contrast and post-contrast image pairs.
We demonstrate the effectiveness of our approach on synthetic and real datasets using various scanners, field strengths, and contrast agents.
arXiv Detail & Related papers (2024-03-06T08:35:29Z) - Forgery-aware Adaptive Transformer for Generalizable Synthetic Image
Detection [106.39544368711427]
We study the problem of generalizable synthetic image detection, aiming to detect forgery images from diverse generative methods.
We present a novel forgery-aware adaptive transformer approach, namely FatFormer.
Our approach tuned on 4-class ProGAN data attains an average of 98% accuracy to unseen GANs, and surprisingly generalizes to unseen diffusion models with 95% accuracy.
arXiv Detail & Related papers (2023-12-27T17:36:32Z) - Synthetic optical coherence tomography angiographs for detailed retinal
vessel segmentation without human annotations [12.571349114534597]
We present a lightweight simulation of the retinal vascular network based on space colonization for faster and more realistic OCTA synthesis.
We demonstrate the superior segmentation performance of our approach in extensive quantitative and qualitative experiments on three public datasets.
arXiv Detail & Related papers (2023-06-19T14:01:47Z) - Spatial gradient consistency for unsupervised learning of hyperspectral
demosaicking: Application to surgical imaging [4.795951381086172]
Hyperspectral imaging has the potential to improve tissue characterisation in real-time and with high-resolution.
A demosaicking algorithm is required to fully recover the spatial and spectral information of the snapshot images.
We present a fully unsupervised hyperspectral image demosaicking algorithm which only requires snapshot images for training purposes.
arXiv Detail & Related papers (2023-02-21T18:07:14Z) - High-fidelity Direct Contrast Synthesis from Magnetic Resonance
Fingerprinting [28.702553164811473]
We propose a supervised learning-based method that directly synthesizes contrast-weighted images from the MRF data without going through the quantitative mapping and spin-dynamics simulation.
In-vivo experiments demonstrate excellent image quality compared to simulation-based contrast synthesis and previous DCS methods, both visually as well as by quantitative metrics.
arXiv Detail & Related papers (2022-12-21T07:11:39Z) - Treatment Learning Causal Transformer for Noisy Image Classification [62.639851972495094]
In this work, we incorporate this binary information of "existence of noise" as treatment into image classification tasks to improve prediction accuracy.
Motivated from causal variational inference, we propose a transformer-based architecture, that uses a latent generative model to estimate robust feature representations for noise image classification.
We also create new noisy image datasets incorporating a wide range of noise factors for performance benchmarking.
arXiv Detail & Related papers (2022-03-29T13:07:53Z) - InDuDoNet+: A Model-Driven Interpretable Dual Domain Network for Metal
Artifact Reduction in CT Images [53.4351366246531]
We construct a novel interpretable dual domain network, termed InDuDoNet+, into which CT imaging process is finely embedded.
We analyze the CT values among different tissues, and merge the prior observations into a prior network for our InDuDoNet+, which significantly improve its generalization performance.
arXiv Detail & Related papers (2021-12-23T15:52:37Z) - Incremental Cross-view Mutual Distillation for Self-supervised Medical
CT Synthesis [88.39466012709205]
This paper builds a novel medical slice to increase the between-slice resolution.
Considering that the ground-truth intermediate medical slices are always absent in clinical practice, we introduce the incremental cross-view mutual distillation strategy.
Our method outperforms state-of-the-art algorithms by clear margins.
arXiv Detail & Related papers (2021-12-20T03:38:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.