Behind the Machine's Gaze: Biologically Constrained Neural Networks
Exhibit Human-like Visual Attention
- URL: http://arxiv.org/abs/2204.09093v1
- Date: Tue, 19 Apr 2022 18:57:47 GMT
- Title: Behind the Machine's Gaze: Biologically Constrained Neural Networks
Exhibit Human-like Visual Attention
- Authors: Leo Schwinn, Doina Precup, Bj\"orn Eskofier, and Dario Zanca
- Abstract summary: We propose the Neural Visual Attention (NeVA) algorithm to generate visual scanpaths in a top-down manner.
We show that the proposed method outperforms state-of-the-art unsupervised human attention models in terms of similarity to human scanpaths.
- Score: 40.878963450471026
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: By and large, existing computational models of visual attention tacitly
assume perfect vision and full access to the stimulus and thereby deviate from
foveated biological vision. Moreover, modelling top-down attention is generally
reduced to the integration of semantic features without incorporating the
signal of a high-level visual tasks that have shown to partially guide human
attention. We propose the Neural Visual Attention (NeVA) algorithm to generate
visual scanpaths in a top-down manner. With our method, we explore the ability
of neural networks on which we impose the biological constraints of foveated
vision to generate human-like scanpaths. Thereby, the scanpaths are generated
to maximize the performance with respect to the underlying visual task (i.e.,
classification or reconstruction). Extensive experiments show that the proposed
method outperforms state-of-the-art unsupervised human attention models in
terms of similarity to human scanpaths. Additionally, the flexibility of the
framework allows to quantitatively investigate the role of different tasks in
the generated visual behaviours. Finally, we demonstrate the superiority of the
approach in a novel experiment that investigates the utility of scanpaths in
real-world applications, where imperfect viewing conditions are given.
Related papers
- Brain Mapping with Dense Features: Grounding Cortical Semantic Selectivity in Natural Images With Vision Transformers [5.265058307999745]
We introduce BrainSAIL, a method for isolating neurally-activating visual concepts in images.
BrainSAIL exploits semantically consistent, dense spatial features from pre-trained vision models.
We validate BrainSAIL on cortical regions with known category selectivity.
arXiv Detail & Related papers (2024-10-07T17:59:45Z) - Unidirectional brain-computer interface: Artificial neural network
encoding natural images to fMRI response in the visual cortex [12.1427193917406]
We propose an artificial neural network dubbed VISION to mimic the human brain and show how it can foster neuroscientific inquiries.
VISION successfully predicts human hemodynamic responses as fMRI voxel values to visual inputs with an accuracy exceeding state-of-the-art performance by 45%.
arXiv Detail & Related papers (2023-09-26T15:38:26Z) - Contrastive Language-Image Pretrained Models are Zero-Shot Human
Scanpath Predictors [2.524526956420465]
CapMIT1003 is a database of captions and click-contingent image explorations collected during captioning tasks.
NevaClip is a novel zero-shot method for predicting visual scanpaths.
arXiv Detail & Related papers (2023-05-21T07:24:50Z) - Simulating Human Gaze with Neural Visual Attention [44.65733084492857]
We propose the Neural Visual Attention (NeVA) algorithm to integrate guidance of any downstream visual task into attention modeling.
We observe that biologically constrained neural networks generate human-like scanpaths without being trained for this objective.
arXiv Detail & Related papers (2022-11-22T09:02:09Z) - An Inter-observer consistent deep adversarial training for visual
scanpath prediction [66.46953851227454]
We propose an inter-observer consistent adversarial training approach for scanpath prediction through a lightweight deep neural network.
We show the competitiveness of our approach in regard to state-of-the-art methods.
arXiv Detail & Related papers (2022-11-14T13:22:29Z) - BI AVAN: Brain inspired Adversarial Visual Attention Network [67.05560966998559]
We propose a brain-inspired adversarial visual attention network (BI-AVAN) to characterize human visual attention directly from functional brain activity.
Our model imitates the biased competition process between attention-related/neglected objects to identify and locate the visual objects in a movie frame the human brain focuses on in an unsupervised manner.
arXiv Detail & Related papers (2022-10-27T22:20:36Z) - Guiding Visual Attention in Deep Convolutional Neural Networks Based on
Human Eye Movements [0.0]
Deep Convolutional Neural Networks (DCNNs) were originally inspired by principles of biological vision.
Recent advances in deep learning seem to decrease this similarity.
We investigate a purely data-driven approach to obtain useful models.
arXiv Detail & Related papers (2022-06-21T17:59:23Z) - Peripheral Vision Transformer [52.55309200601883]
We take a biologically inspired approach and explore to model peripheral vision in deep neural networks for visual recognition.
We propose to incorporate peripheral position encoding to the multi-head self-attention layers to let the network learn to partition the visual field into diverse peripheral regions given training data.
We evaluate the proposed network, dubbed PerViT, on the large-scale ImageNet dataset and systematically investigate the inner workings of the model for machine perception.
arXiv Detail & Related papers (2022-06-14T12:47:47Z) - FuNNscope: Visual microscope for interactively exploring the loss
landscape of fully connected neural networks [77.34726150561087]
We show how to explore high-dimensional landscape characteristics of neural networks.
We generalize observations on small neural networks to more complex systems.
An interactive dashboard opens up a number of possible application networks.
arXiv Detail & Related papers (2022-04-09T16:41:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.