A Logic of General Attention Using Edge-Conditioned Event Models (Extended Version)
- URL: http://arxiv.org/abs/2505.14539v1
- Date: Tue, 20 May 2025 15:56:34 GMT
- Title: A Logic of General Attention Using Edge-Conditioned Event Models (Extended Version)
- Authors: Gaia Belardinelli, Thomas Bolander, Sebastian Watzl,
- Abstract summary: We present the first general logic of attention.<n>Our work treats attention as a modality, like belief or awareness.<n>We illustrate our framework with examples of AI agents reasoning about human attentional biases.
- Score: 1.6199400106794555
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we present the first general logic of attention. Attention is a powerful cognitive ability that allows agents to focus on potentially complex information, such as logically structured propositions, higher-order beliefs, or what other agents pay attention to. This ability is a strength, as it helps to ignore what is irrelevant, but it can also introduce biases when some types of information or agents are systematically ignored. Existing dynamic epistemic logics for attention cannot model such complex attention scenarios, as they only model attention to atomic formulas. Additionally, such logics quickly become cumbersome, as their size grows exponentially in the number of agents and announced literals. Here, we introduce a logic that overcomes both limitations. First, we generalize edge-conditioned event models, which we show to be as expressive as standard event models yet exponentially more succinct (generalizing both standard event models and generalized arrow updates). Second, we extend attention to arbitrary formulas, allowing agents to also attend to other agents' beliefs or attention. Our work treats attention as a modality, like belief or awareness. We introduce attention principles that impose closure properties on that modality and that can be used in its axiomatization. Throughout, we illustrate our framework with examples of AI agents reasoning about human attentional biases, demonstrating how such agents can discover attentional biases.
Related papers
- When Thinking Backfires: Mechanistic Insights Into Reasoning-Induced Misalignment [23.096167213579957]
Reasoning-Induced Misalignment (RIM) emerges when reasoning capabilities strengthen.<n>RIM is caused when specific types of reasoning patterns are introduced during inference or training.<n>During training, we find significantly higher activation entanglement between reasoning and safety in safety-critical neurons.
arXiv Detail & Related papers (2025-08-30T16:04:54Z) - Transformers Learn Faster with Semantic Focus [57.97235825738412]
We study sparse transformers in terms of learnability and generalization.<n>We find that input-dependent sparse attention models appear to converge faster and generalize better than standard attention models.
arXiv Detail & Related papers (2025-06-17T01:19:28Z) - Testing Components of the Attention Schema Theory in Artificial Neural Networks [0.0]
Growing evidence suggests that the brain uses an attention schema, or a simplified model of attention, to help control what it attends to.<n>The effects of an attention schema may be examined in artificial agents.
arXiv Detail & Related papers (2024-11-01T19:18:07Z) - Agent Attention: On the Integration of Softmax and Linear Attention [70.06472039237354]
We propose a novel attention paradigm, Agent Attention, to strike a favorable balance between computational efficiency and representation power.
We show that the proposed agent attention is equivalent to a generalized form of linear attention.
Notably, agent attention has shown remarkable performance in high-resolution scenarios, owning to its linear attention nature.
arXiv Detail & Related papers (2023-12-14T16:26:29Z) - Attention Schema in Neural Agents [66.43628974353683]
In cognitive neuroscience, Attention Theory (AST) supports the idea of distinguishing attention from AS.
AST predicts that an agent can use its own AS to also infer the states of other agents' attention.
We explore different ways in which attention and AS interact with each other.
arXiv Detail & Related papers (2023-05-27T05:40:34Z) - Attention! Dynamic Epistemic Logic Models of (In)attentive Agents [3.6933317368929197]
We propose a generalization that allows for paying attention to subsets of atomic formulas.
We then extend the framework to account for inattentive agents that, instead of assuming nothing happens, may default to a specific truth-value.
arXiv Detail & Related papers (2023-03-23T17:55:32Z) - Revisiting Attention Weights as Explanations from an Information
Theoretic Perspective [4.499369811647602]
We show that attention mechanisms have the potential to function as a shortcut to model explanations when they are carefully combined with other model elements.
Our findings indicate that attention mechanisms do have the potential to function as a shortcut to model explanations when they are carefully combined with other model elements.
arXiv Detail & Related papers (2022-10-31T12:53:20Z) - Guiding Visual Question Answering with Attention Priors [76.21671164766073]
We propose to guide the attention mechanism using explicit linguistic-visual grounding.
This grounding is derived by connecting structured linguistic concepts in the query to their referents among the visual objects.
The resultant algorithm is capable of probing attention-based reasoning models, injecting relevant associative knowledge, and regulating the core reasoning process.
arXiv Detail & Related papers (2022-05-25T09:53:47Z) - Is Sparse Attention more Interpretable? [52.85910570651047]
We investigate how sparsity affects our ability to use attention as an explainability tool.
We find that only a weak relationship between inputs and co-indexed intermediate representations exists -- under sparse attention.
We observe in this setting that inducing sparsity may make it less plausible that attention can be used as a tool for understanding model behavior.
arXiv Detail & Related papers (2021-06-02T11:42:56Z) - Abstract Spatial-Temporal Reasoning via Probabilistic Abduction and
Execution [97.50813120600026]
Spatial-temporal reasoning is a challenging task in Artificial Intelligence (AI)
Recent works have focused on an abstract reasoning task of this kind -- Raven's Progressive Matrices ( RPM)
We propose a neuro-symbolic Probabilistic Abduction and Execution learner (PrAE) learner.
arXiv Detail & Related papers (2021-03-26T02:42:18Z) - Attention or memory? Neurointerpretable agents in space and time [0.0]
We design a model incorporating a self-attention mechanism that implements task-state representations in semantic feature-space.
To evaluate the agent's selective properties, we add a large volume of task-irrelevant features to observations.
In line with neuroscience predictions, self-attention leads to increased robustness to noise compared to benchmark models.
arXiv Detail & Related papers (2020-07-09T15:04:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.