How attention simplifies mental representations for planning
- URL: http://arxiv.org/abs/2506.09520v1
- Date: Wed, 11 Jun 2025 08:46:05 GMT
- Title: How attention simplifies mental representations for planning
- Authors: Jason da Silva Castanheira, Nicholas Shea, Stephen M. Fleming,
- Abstract summary: We study how spatial attention controls which aspects of a task representation enter subjective awareness and are available for planning.<n>We find that spatial proximity governs which aspects of a maze are available for planning.<n>Inspired by the'spotlight of attention' analogy, we incorporate the effects of visuospatial attention into existing computational accounts of value-guided construal.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Human planning is efficient -- it frugally deploys limited cognitive resources to accomplish difficult tasks -- and flexible -- adapting to novel problems and environments. Computational approaches suggest that people construct simplified mental representations of their environment, balancing the complexity of a task representation with its utility. These models imply a nested optimisation in which planning shapes perception, and perception shapes planning -- but the perceptual and attentional mechanisms governing how this interaction unfolds remain unknown. Here, we harness virtual maze navigation to characterise how spatial attention controls which aspects of a task representation enter subjective awareness and are available for planning. We find that spatial proximity governs which aspects of a maze are available for planning, and that when task-relevant information follows natural (lateralised) contours of attention, people can more easily construct simplified and useful maze representations. This influence of attention varies considerably across individuals, explaining differences in people's task representations and behaviour. Inspired by the 'spotlight of attention' analogy, we incorporate the effects of visuospatial attention into existing computational accounts of value-guided construal. Together, our work bridges computational perspectives on perception and decision-making to better understand how individuals represent their environments in aid of planning.
Related papers
- Visual Environment-Interactive Planning for Embodied Complex-Question Answering [28.929345360469807]
This study focuses on Embodied Complex-Question Answering task.<n>The core of this task lies in making appropriate plans based on the perception of the visual environment.<n>Considering multi-step planning, the framework for formulating plans in a sequential manner is proposed in this paper.
arXiv Detail & Related papers (2025-04-01T13:26:28Z) - Temporal Representation Alignment: Successor Features Enable Emergent Compositionality in Robot Instruction Following [50.377287115281476]
We show that learning to associate the representations of current and future states with a temporal loss can improve compositional generalization.<n>We evaluate our approach across diverse robotic manipulation tasks as well as in simulation, showing substantial improvements for tasks specified with either language or goal images.
arXiv Detail & Related papers (2025-02-08T05:26:29Z) - PlanCritic: Formal Planning with Human Feedback [0.994853090657971]
Real world planning problems are often too complex to be effectively tackled by a single unaided human.<n>Recent work has focused on developing a collaborative planning system to assist humans in complex domains.<n>We present a feedback-driven plan critic as part of a broader cooperative planning system.
arXiv Detail & Related papers (2024-11-30T00:58:48Z) - VSP: Assessing the dual challenges of perception and reasoning in spatial planning tasks for VLMs [102.36953558562436]
Vision language models (VLMs) are an exciting emerging class of language models (LMs)
One understudied capability inVLMs is visual spatial planning.
Our study introduces a benchmark that evaluates the spatial planning capability in these models in general.
arXiv Detail & Related papers (2024-07-02T00:24:01Z) - Self-Explainable Affordance Learning with Embodied Caption [63.88435741872204]
We introduce Self-Explainable Affordance learning (SEA) with embodied caption.
SEA enables robots to articulate their intentions and bridge the gap between explainable vision-language caption and visual affordance learning.
We propose a novel model to effectively combine affordance grounding with self-explanation in a simple but efficient manner.
arXiv Detail & Related papers (2024-04-08T15:22:38Z) - Dynamic planning in hierarchical active inference [0.0]
We refer to the ability of the human brain to infer and impose motor trajectories related to cognitive decisions.
This study focuses on the topic of dynamic planning in active inference.
arXiv Detail & Related papers (2024-02-18T17:32:53Z) - InstructDiffusion: A Generalist Modeling Interface for Vision Tasks [52.981128371910266]
We present InstructDiffusion, a framework for aligning computer vision tasks with human instructions.
InstructDiffusion could handle a variety of vision tasks, including understanding tasks and generative tasks.
It even exhibits the ability to handle unseen tasks and outperforms prior methods on novel datasets.
arXiv Detail & Related papers (2023-09-07T17:56:57Z) - Attention Mechanisms in Computer Vision: A Survey [75.6074182122423]
We provide a comprehensive review of various attention mechanisms in computer vision.
We categorize them according to approach, such as channel attention, spatial attention, temporal attention and branch attention.
We suggest future directions for attention mechanism research.
arXiv Detail & Related papers (2021-11-15T09:18:40Z) - Understanding top-down attention using task-oriented ablation design [0.22940141855172028]
Top-down attention allows neural networks, both artificial and biological, to focus on the information most relevant for a given task.
We aim to answer this with a computational experiment based on a general framework called task-oriented ablation design.
We compare the performance of two neural networks, one with top-down attention and one without.
arXiv Detail & Related papers (2021-06-08T21:01:47Z) - Control of mental representations in human planning [38.227123320091046]
Two aspects of human planning stand out its efficiency, even in complex environments, and its flexibility, even in changing environments.
Efficiency is especially impressive because directly computing an optimal plan is intractable, even for modestly complex tasks, and yet people successfully solve myriad problems despite limited cognitive resources.
Here, we propose that mental representations can be controlled and that this provides opportunities to adaptively simplify problems so they can be more easily reasoned about.
arXiv Detail & Related papers (2021-05-14T16:39:31Z) - Learning Obstacle Representations for Neural Motion Planning [70.80176920087136]
We address sensor-based motion planning from a learning perspective.
Motivated by recent advances in visual recognition, we argue the importance of learning appropriate representations for motion planning.
We propose a new obstacle representation based on the PointNet architecture and train it jointly with policies for obstacle avoidance.
arXiv Detail & Related papers (2020-08-25T17:12:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.