Deep Learning-Based Perceptual Stimulus Encoder for Bionic Vision
- URL: http://arxiv.org/abs/2203.05604v1
- Date: Thu, 10 Mar 2022 19:42:09 GMT
- Title: Deep Learning-Based Perceptual Stimulus Encoder for Bionic Vision
- Authors: Lucas Relic, Bowen Zhang, Yi-Lin Tuan, Michael Beyeler
- Abstract summary: We propose a PSE that is trained in an end-to-end fashion to predict the electrode activation patterns required to produce a desired visual percept.
We demonstrate the effectiveness of the encoder on MNIST using a psychophysically validated phosphene model tailored to individual retinal implant users.
- Score: 6.1739856715198
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Retinal implants have the potential to treat incurable blindness, yet the
quality of the artificial vision they produce is still rudimentary. An
outstanding challenge is identifying electrode activation patterns that lead to
intelligible visual percepts (phosphenes). Here we propose a PSE based on CNN
that is trained in an end-to-end fashion to predict the electrode activation
patterns required to produce a desired visual percept. We demonstrate the
effectiveness of the encoder on MNIST using a psychophysically validated
phosphene model tailored to individual retinal implant users. The present work
constitutes an essential first step towards improving the quality of the
artificial vision provided by retinal implants.
Related papers
- ReWaRD: Retinal Waves for Pre-Training Artificial Neural Networks
Mimicking Real Prenatal Development [5.222115919729418]
Pre- and post-natal retinal waves suggest to be a pre-training mechanism for the primate visual system.
We build a computational model that mimics this development mechanism by pre-training different artificial convolutional neural networks.
The resulting features of this biologically plausible pre-training closely match the V1 features of the primate visual system.
arXiv Detail & Related papers (2023-11-28T21:14:05Z) - Adapting Brain-Like Neural Networks for Modeling Cortical Visual
Prostheses [68.96380145211093]
Cortical prostheses are devices implanted in the visual cortex that attempt to restore lost vision by electrically stimulating neurons.
Currently, the vision provided by these devices is limited, and accurately predicting the visual percepts resulting from stimulation is an open challenge.
We propose to address this challenge by utilizing 'brain-like' convolutional neural networks (CNNs), which have emerged as promising models of the visual system.
arXiv Detail & Related papers (2022-09-27T17:33:19Z) - Towards Trustworthy Healthcare AI: Attention-Based Feature Learning for
COVID-19 Screening With Chest Radiography [70.37371604119826]
Building AI models with trustworthiness is important especially in regulated areas such as healthcare.
Previous work uses convolutional neural networks as the backbone architecture, which has shown to be prone to over-caution and overconfidence in making decisions.
We propose a feature learning approach using Vision Transformers, which use an attention-based mechanism.
arXiv Detail & Related papers (2022-07-19T14:55:42Z) - Peripheral Vision Transformer [52.55309200601883]
We take a biologically inspired approach and explore to model peripheral vision in deep neural networks for visual recognition.
We propose to incorporate peripheral position encoding to the multi-head self-attention layers to let the network learn to partition the visual field into diverse peripheral regions given training data.
We evaluate the proposed network, dubbed PerViT, on the large-scale ImageNet dataset and systematically investigate the inner workings of the model for machine perception.
arXiv Detail & Related papers (2022-06-14T12:47:47Z) - Cross-modal Clinical Graph Transformer for Ophthalmic Report Generation [116.87918100031153]
We propose a Cross-modal clinical Graph Transformer (CGT) for ophthalmic report generation (ORG)
CGT injects clinical relation triples into the visual features as prior knowledge to drive the decoding procedure.
Experiments on the large-scale FFA-IR benchmark demonstrate that the proposed CGT is able to outperform previous benchmark methods.
arXiv Detail & Related papers (2022-06-04T13:16:30Z) - A Hybrid Neural Autoencoder for Sensory Neuroprostheses and Its
Applications in Bionic Vision [0.0]
Sensory neuroprostheses are emerging as a promising technology to restore lost sensory function or augment human capacities.
In this paper we show how a deep neural network encoder is trained to invert a known, fixed forward model that approximates the underlying biological system.
As a proof of concept, we demonstrate the effectiveness of our hybrid neural autoencoder (HNA) on the use case of visual neuroprostheses.
arXiv Detail & Related papers (2022-05-26T20:52:00Z) - Efficient visual object representation using a biologically plausible
spike-latency code and winner-take-all inhibition [0.0]
spiking neural networks (SNNs) have the potential to improve the efficiency and biological plausibility of object recognition systems.
We present a SNN model that uses spike-latency coding and winner-take-all inhibition (WTA-I) to efficiently represent visual stimuli.
We demonstrate that a network of 150 spiking neurons can efficiently represent objects with as little as 40 spikes.
arXiv Detail & Related papers (2022-05-20T17:48:02Z) - An Interpretable Multiple-Instance Approach for the Detection of
referable Diabetic Retinopathy from Fundus Images [72.94446225783697]
We propose a machine learning system for the detection of referable Diabetic Retinopathy in fundus images.
By extracting local information from image patches and combining it efficiently through an attention mechanism, our system is able to achieve high classification accuracy.
We evaluate our approach on publicly available retinal image datasets, in which it exhibits near state-of-the-art performance.
arXiv Detail & Related papers (2021-03-02T13:14:15Z) - Deep Learning--Based Scene Simplification for Bionic Vision [0.0]
We show that object segmentation may better support scene understanding than models based on visual saliency and monocular depth estimation.
This work has the potential to drastically improve the utility of prosthetic vision for people blinded from retinal degenerative diseases.
arXiv Detail & Related papers (2021-01-30T19:35:33Z) - Retinopathy of Prematurity Stage Diagnosis Using Object Segmentation and
Convolutional Neural Networks [68.96150598294072]
Retinopathy of Prematurity (ROP) is an eye disorder primarily affecting premature infants with lower weights.
It causes proliferation of vessels in the retina and could result in vision loss and, eventually, retinal detachment, leading to blindness.
In recent years, there has been a significant effort to automate the diagnosis using deep learning.
This paper builds upon the success of previous models and develops a novel architecture, which combines object segmentation and convolutional neural networks (CNN)
Our proposed system first trains an object segmentation model to identify the demarcation line at a pixel level and adds the resulting mask as an additional "color" channel in
arXiv Detail & Related papers (2020-04-03T14:07:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.