The Brain-Inspired Decoder for Natural Visual Image Reconstruction
- URL: http://arxiv.org/abs/2207.08591v1
- Date: Mon, 18 Jul 2022 13:31:26 GMT
- Title: The Brain-Inspired Decoder for Natural Visual Image Reconstruction
- Authors: Wenyi Li, Shengjie Zheng, Yufan Liao, Rongqi Hong, Weiliang Chen,
Chenggnag He, Xiaojian Li
- Abstract summary: We propose a deep learning neural network architecture with biological properties to reconstruct visual image from spike trains.
Our model is an end-to-end decoder from neural spike trains to images.
Our results show that our method can effectively combine receptive field features to reconstruct images.
- Score: 4.433315630787158
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Decoding images from brain activity has been a challenge. Owing to the
development of deep learning, there are available tools to solve this problem.
The decoded image, which aims to map neural spike trains to low-level visual
features and high-level semantic information space. Recently, there are a few
studies of decoding from spike trains, however, these studies pay less
attention to the foundations of neuroscience and there are few studies that
merged receptive field into visual image reconstruction. In this paper, we
propose a deep learning neural network architecture with biological properties
to reconstruct visual image from spike trains. As far as we know, we
implemented a method that integrated receptive field property matrix into loss
function at the first time. Our model is an end-to-end decoder from neural
spike trains to images. We not only merged Gabor filter into auto-encoder which
used to generate images but also proposed a loss function with receptive field
properties. We evaluated our decoder on two datasets which contain macaque
primary visual cortex neural spikes and salamander retina ganglion cells (RGCs)
spikes. Our results show that our method can effectively combine receptive
field features to reconstruct images, providing a new approach to visual
reconstruction based on neural information.
Related papers
- Learning Multimodal Volumetric Features for Large-Scale Neuron Tracing [72.45257414889478]
We aim to reduce human workload by predicting connectivity between over-segmented neuron pieces.
We first construct a dataset, named FlyTracing, that contains millions of pairwise connections of segments expanding the whole fly brain.
We propose a novel connectivity-aware contrastive learning method to generate dense volumetric EM image embedding.
arXiv Detail & Related papers (2024-01-05T19:45:12Z) - Improving visual image reconstruction from human brain activity using
latent diffusion models via multiple decoded inputs [2.4366811507669124]
Integration of deep learning and neuroscience has led to improvements in the analysis of brain activity.
The reconstruction of visual experience from human brain activity is an area that has particularly benefited.
We examine the extent to which various additional decoding techniques affect the performance of visual experience reconstruction.
arXiv Detail & Related papers (2023-06-20T13:48:02Z) - Brain Captioning: Decoding human brain activity into images and text [1.5486926490986461]
We present an innovative method for decoding brain activity into meaningful images and captions.
Our approach takes advantage of cutting-edge image captioning models and incorporates a unique image reconstruction pipeline.
We evaluate our methods using quantitative metrics for both generated captions and images.
arXiv Detail & Related papers (2023-05-19T09:57:19Z) - Mind Reader: Reconstructing complex images from brain activities [16.78619734818198]
We focus on reconstructing the complex image stimuli from fMRI (functional magnetic resonance imaging) signals.
Unlike previous works that reconstruct images with single objects or simple shapes, our work aims to reconstruct image stimuli rich in semantics.
We find that incorporating an additional text modality is beneficial for the reconstruction problem compared to directly translating brain signals to images.
arXiv Detail & Related papers (2022-09-30T06:32:46Z) - Attentive Symmetric Autoencoder for Brain MRI Segmentation [56.02577247523737]
We propose a novel Attentive Symmetric Auto-encoder based on Vision Transformer (ViT) for 3D brain MRI segmentation tasks.
In the pre-training stage, the proposed auto-encoder pays more attention to reconstruct the informative patches according to the gradient metrics.
Experimental results show that our proposed attentive symmetric auto-encoder outperforms the state-of-the-art self-supervised learning methods and medical image segmentation models.
arXiv Detail & Related papers (2022-09-19T09:43:19Z) - Is Deep Image Prior in Need of a Good Education? [57.3399060347311]
Deep image prior was introduced as an effective prior for image reconstruction.
Despite its impressive reconstructive properties, the approach is slow when compared to learned or traditional reconstruction techniques.
We develop a two-stage learning paradigm to address the computational challenge.
arXiv Detail & Related papers (2021-11-23T15:08:26Z) - Neural Photofit: Gaze-based Mental Image Reconstruction [25.67771238116104]
We propose a novel method that leverages human fixations to visually decode the image a person has in mind into a photofit (facial composite)
Our method combines three neural networks: An encoder, a scoring network, and a decoder.
We show that our method significantly outperforms a mean baseline predictor and report on a human study that shows that we can decode photofits that are visually plausible and close to the observer's mental image.
arXiv Detail & Related papers (2021-08-17T09:11:32Z) - Controlled Caption Generation for Images Through Adversarial Attacks [85.66266989600572]
We study adversarial examples for vision and language models, which typically adopt a Convolutional Neural Network (i.e., CNN) for image feature extraction and a Recurrent Neural Network (RNN) for caption generation.
In particular, we investigate attacks on the visual encoder's hidden layer that is fed to the subsequent recurrent network.
We propose a GAN-based algorithm for crafting adversarial examples for neural image captioning that mimics the internal representation of the CNN.
arXiv Detail & Related papers (2021-07-07T07:22:41Z) - NAS-DIP: Learning Deep Image Prior with Neural Architecture Search [65.79109790446257]
Recent work has shown that the structure of deep convolutional neural networks can be used as a structured image prior.
We propose to search for neural architectures that capture stronger image priors.
We search for an improved network by leveraging an existing neural architecture search algorithm.
arXiv Detail & Related papers (2020-08-26T17:59:36Z) - Neural Sparse Representation for Image Restoration [116.72107034624344]
Inspired by the robustness and efficiency of sparse coding based image restoration models, we investigate the sparsity of neurons in deep networks.
Our method structurally enforces sparsity constraints upon hidden neurons.
Experiments show that sparse representation is crucial in deep neural networks for multiple image restoration tasks.
arXiv Detail & Related papers (2020-06-08T05:15:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.