BI AVAN: Brain inspired Adversarial Visual Attention Network
- URL: http://arxiv.org/abs/2210.15790v1
- Date: Thu, 27 Oct 2022 22:20:36 GMT
- Title: BI AVAN: Brain inspired Adversarial Visual Attention Network
- Authors: Heng Huang, Lin Zhao, Xintao Hu, Haixing Dai, Lu Zhang, Dajiang Zhu,
Tianming Liu
- Abstract summary: We propose a brain-inspired adversarial visual attention network (BI-AVAN) to characterize human visual attention directly from functional brain activity.
Our model imitates the biased competition process between attention-related/neglected objects to identify and locate the visual objects in a movie frame the human brain focuses on in an unsupervised manner.
- Score: 67.05560966998559
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual attention is a fundamental mechanism in the human brain, and it
inspires the design of attention mechanisms in deep neural networks. However,
most of the visual attention studies adopted eye-tracking data rather than the
direct measurement of brain activity to characterize human visual attention. In
addition, the adversarial relationship between the attention-related objects
and attention-neglected background in the human visual system was not fully
exploited. To bridge these gaps, we propose a novel brain-inspired adversarial
visual attention network (BI-AVAN) to characterize human visual attention
directly from functional brain activity. Our BI-AVAN model imitates the biased
competition process between attention-related/neglected objects to identify and
locate the visual objects in a movie frame the human brain focuses on in an
unsupervised manner. We use independent eye-tracking data as ground truth for
validation and experimental results show that our model achieves robust and
promising results when inferring meaningful human visual attention and mapping
the relationship between brain activities and visual stimuli. Our BI-AVAN model
contributes to the emerging field of leveraging the brain's functional
architecture to inspire and guide the model design in artificial intelligence
(AI), e.g., deep neural networks.
Related papers
- BRACTIVE: A Brain Activation Approach to Human Visual Brain Learning [11.517021103782229]
We introduce Brain Activation Network (BRACTIVE), a transformer-based approach to studying the human visual brain.
The main objective of BRACTIVE is to align the visual features of subjects with corresponding brain representations via fMRI signals.
Our experiments demonstrate that BRACTIVE effectively identifies person-specific regions of interest, such as face and body-selective areas.
arXiv Detail & Related papers (2024-05-29T06:50:13Z) - Achieving More Human Brain-Like Vision via Human EEG Representational Alignment [1.811217832697894]
We present 'Re(presentational)Al(ignment)net', a vision model aligned with human brain activity based on non-invasive EEG.
Our innovative image-to-brain multi-layer encoding framework advances human neural alignment by optimizing multiple model layers.
Our findings suggest that ReAlnet represents a breakthrough in bridging the gap between artificial and human vision, and paving the way for more brain-like artificial intelligence systems.
arXiv Detail & Related papers (2024-01-30T18:18:41Z) - Adapting Brain-Like Neural Networks for Modeling Cortical Visual
Prostheses [68.96380145211093]
Cortical prostheses are devices implanted in the visual cortex that attempt to restore lost vision by electrically stimulating neurons.
Currently, the vision provided by these devices is limited, and accurately predicting the visual percepts resulting from stimulation is an open challenge.
We propose to address this challenge by utilizing 'brain-like' convolutional neural networks (CNNs), which have emerged as promising models of the visual system.
arXiv Detail & Related papers (2022-09-27T17:33:19Z) - Human Eyes Inspired Recurrent Neural Networks are More Robust Against Adversarial Noises [7.689542442882423]
We designed a dual-stream vision model inspired by the human brain.
This model features retina-like input layers and includes two streams: one determining the next point of focus (the fixation), while the other interprets the visuals surrounding the fixation.
We evaluated this model against various benchmarks in terms of object recognition, gaze behavior and adversarial robustness.
arXiv Detail & Related papers (2022-06-15T03:44:42Z) - Functional2Structural: Cross-Modality Brain Networks Representation
Learning [55.24969686433101]
Graph mining on brain networks may facilitate the discovery of novel biomarkers for clinical phenotypes and neurodegenerative diseases.
We propose a novel graph learning framework, known as Deep Signed Brain Networks (DSBN), with a signed graph encoder.
We validate our framework on clinical phenotype and neurodegenerative disease prediction tasks using two independent, publicly available datasets.
arXiv Detail & Related papers (2022-05-06T03:45:36Z) - Attention Mechanisms in Computer Vision: A Survey [75.6074182122423]
We provide a comprehensive review of various attention mechanisms in computer vision.
We categorize them according to approach, such as channel attention, spatial attention, temporal attention and branch attention.
We suggest future directions for attention mechanism research.
arXiv Detail & Related papers (2021-11-15T09:18:40Z) - SOLVER: Scene-Object Interrelated Visual Emotion Reasoning Network [83.27291945217424]
We propose a novel Scene-Object interreLated Visual Emotion Reasoning network (SOLVER) to predict emotions from images.
To mine the emotional relationships between distinct objects, we first build up an Emotion Graph based on semantic concepts and visual features.
We also design a Scene-Object Fusion Module to integrate scenes and objects, which exploits scene features to guide the fusion process of object features with the proposed scene-based attention mechanism.
arXiv Detail & Related papers (2021-10-24T02:41:41Z) - Object Based Attention Through Internal Gating [4.941630596191806]
We propose an artificial neural network model of object-based attention.
Our model captures the way in which attention is both top-down and recurrent.
We find that our model replicates a range of findings from neuroscience.
arXiv Detail & Related papers (2021-06-08T17:20:50Z) - Brain-inspired algorithms for processing of visual data [5.045960549713147]
We review approaches for image processing and computer vision based on neuro-scientific findings about the functions of some neurons in the visual cortex.
We pay particular attention to the mechanisms of inhibition of the responses of some neurons, which provide the visual system with improved stability to changing input stimuli.
arXiv Detail & Related papers (2021-03-02T10:45:38Z) - Visual Relationship Detection with Visual-Linguistic Knowledge from
Multimodal Representations [103.00383924074585]
Visual relationship detection aims to reason over relationships among salient objects in images.
We propose a novel approach named Visual-Linguistic Representations from Transformers (RVL-BERT)
RVL-BERT performs spatial reasoning with both visual and language commonsense knowledge learned via self-supervised pre-training.
arXiv Detail & Related papers (2020-09-10T16:15:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.