The perceptual boost of visual attention is task-dependent in
naturalistic settings
- URL: http://arxiv.org/abs/2003.00882v2
- Date: Mon, 6 Apr 2020 14:30:14 GMT
- Title: The perceptual boost of visual attention is task-dependent in
naturalistic settings
- Authors: Freddie Bickford Smith, Xiaoliang Luo, Brett D. Roads, Bradley C. Love
- Abstract summary: We design a collection of visual tasks, each consisting of classifying images from a chosen task set.
The nature of a task is determined by which categories are included in the task set.
On each task we train an attention-augmented neural network and then compare its accuracy to that of a baseline network.
We show that the perceptual boost of attention is stronger with increasing task-set difficulty, weaker with increasing task-set size and weaker with increasing perceptual similarity within a task set.
- Score: 5.735035463793008
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Top-down attention allows people to focus on task-relevant visual
information. Is the resulting perceptual boost task-dependent in naturalistic
settings? We aim to answer this with a large-scale computational experiment.
First, we design a collection of visual tasks, each consisting of classifying
images from a chosen task set (subset of ImageNet categories). The nature of a
task is determined by which categories are included in the task set. Second, on
each task we train an attention-augmented neural network and then compare its
accuracy to that of a baseline network. We show that the perceptual boost of
attention is stronger with increasing task-set difficulty, weaker with
increasing task-set size and weaker with increasing perceptual similarity
within a task set.
Related papers
- Video Task Decathlon: Unifying Image and Video Tasks in Autonomous
Driving [85.62076860189116]
Video Task Decathlon (VTD) includes ten representative image and video tasks spanning classification, segmentation, localization, and association of objects and pixels.
We develop our unified network, VTDNet, that uses a single structure and a single set of weights for all ten tasks.
arXiv Detail & Related papers (2023-09-08T16:33:27Z) - A Dynamic Feature Interaction Framework for Multi-task Visual Perception [100.98434079696268]
We devise an efficient unified framework to solve multiple common perception tasks.
These tasks include instance segmentation, semantic segmentation, monocular 3D detection, and depth estimation.
Our proposed framework, termed D2BNet, demonstrates a unique approach to parameter-efficient predictions for multi-task perception.
arXiv Detail & Related papers (2023-06-08T09:24:46Z) - Top-Down Visual Attention from Analysis by Synthesis [87.47527557366593]
We consider top-down attention from a classic Analysis-by-Synthesis (AbS) perspective of vision.
We propose Analysis-by-Synthesis Vision Transformer (AbSViT), which is a top-down modulated ViT model that variationally approximates AbS, and controllable achieves top-down attention.
arXiv Detail & Related papers (2023-03-23T05:17:05Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z) - Answer-Me: Multi-Task Open-Vocabulary Visual Question Answering [43.07139534653485]
We present Answer-Me, a task-aware multi-task framework.
We pre-train a vision-language joint model, which is multi-task as well.
Results show state-of-the-art performance, zero-shot generalization, robustness to forgetting, and competitive single-task results.
arXiv Detail & Related papers (2022-05-02T14:53:13Z) - Efficiently Identifying Task Groupings for Multi-Task Learning [55.80489920205404]
Multi-task learning can leverage information learned by one task to benefit the training of other tasks.
We suggest an approach to select which tasks should train together in multi-task learning models.
Our method determines task groupings in a single training run by co-training all tasks together and quantifying the effect to which one task's gradient would affect another task's loss.
arXiv Detail & Related papers (2021-09-10T02:01:43Z) - Improving Robot Localisation by Ignoring Visual Distraction [34.8860186009308]
We introduce Neural Blindness, which gives an agent the ability to completely ignore objects or classes that are deemed distractors.
More explicitly, we aim to render a neural network completely incapable of representing specific chosen classes in its latent space.
In a very real sense, this makes the network "blind" to certain classes, allowing and agent to focus on what is important for a given task, and demonstrates how this can be used to improve localisation.
arXiv Detail & Related papers (2021-07-25T17:45:17Z) - Understanding top-down attention using task-oriented ablation design [0.22940141855172028]
Top-down attention allows neural networks, both artificial and biological, to focus on the information most relevant for a given task.
We aim to answer this with a computational experiment based on a general framework called task-oriented ablation design.
We compare the performance of two neural networks, one with top-down attention and one without.
arXiv Detail & Related papers (2021-06-08T21:01:47Z) - CompositeTasking: Understanding Images by Spatial Composition of Tasks [85.95743368954233]
CompositeTasking is the fusion of multiple, spatially distributed tasks.
The proposed network takes a pair of an image and a set of pixel-wise dense tasks as inputs, and makes the task related predictions for each pixel.
It not only offers us a compact network for multi-tasking, but also allows for task-editing.
arXiv Detail & Related papers (2020-12-16T15:47:02Z) - Deep Multi-Task Augmented Feature Learning via Hierarchical Graph Neural
Network [4.121467410954028]
We propose a Hierarchical Graph Neural Network to learn augmented features for deep multi-task learning.
Experiments on real-world datastes show the significant performance improvement when using this strategy.
arXiv Detail & Related papers (2020-02-12T06:02:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.