Attention! Dynamic Epistemic Logic Models of (In)attentive Agents
- URL: http://arxiv.org/abs/2303.13494v2
- Date: Thu, 18 May 2023 13:41:27 GMT
- Title: Attention! Dynamic Epistemic Logic Models of (In)attentive Agents
- Authors: Gaia Belardinelli and Thomas Bolander
- Abstract summary: We propose a generalization that allows for paying attention to subsets of atomic formulas.
We then extend the framework to account for inattentive agents that, instead of assuming nothing happens, may default to a specific truth-value.
- Score: 3.6933317368929197
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Attention is the crucial cognitive ability that limits and selects what
information we observe. Previous work by Bolander et al. (2016) proposes a
model of attention based on dynamic epistemic logic (DEL) where agents are
either fully attentive or not attentive at all. While introducing the realistic
feature that inattentive agents believe nothing happens, the model does not
represent the most essential aspect of attention: its selectivity. Here, we
propose a generalization that allows for paying attention to subsets of atomic
formulas. We introduce the corresponding logic for propositional attention, and
show its axiomatization to be sound and complete. We then extend the framework
to account for inattentive agents that, instead of assuming nothing happens,
may default to a specific truth-value of what they failed to attend to (a sort
of prior concerning the unattended atoms). This feature allows for a more
cognitively plausible representation of the inattentional blindness phenomenon,
where agents end up with false beliefs due to their failure to attend to
conspicuous but unexpected events. Both versions of the model define
attention-based learning through appropriate DEL event models based on a few
and clear edge principles. While the size of such event models grow
exponentially both with the number of agents and the number of atoms, we
introduce a new logical language for describing event models syntactically and
show that using this language our event models can be represented linearly in
the number of agents and atoms. Furthermore, representing our event models
using this language is achieved by a straightforward formalisation of the
aforementioned edge principles.
Related papers
- The Imperfective Paradox in Large Language Models [19.058068907991277]
We investigate the Imperfective Paradox, where the past progressive aspect entails event realization for activities but not for accomplishments.<n>We introduce ImperfectiveNLI, a diagnostic dataset designed to probe this distinction across diverse semantic classes.<n>We uncover a pervasive Teleological Bias: models systematically hallucinate completion for goal-oriented events, often overriding explicit textual negation.
arXiv Detail & Related papers (2026-01-14T10:57:16Z) - Current Agents Fail to Leverage World Model as Tool for Foresight [61.82522354207919]
Generative world models offer a promising remedy: agents could use them to foresee outcomes before acting.<n>This paper empirically examines whether current agents can leverage such world models as tools to enhance their cognition.
arXiv Detail & Related papers (2026-01-07T13:15:23Z) - From Black-box to Causal-box: Towards Building More Interpretable Models [57.23201263629627]
We introduce the notion of causal interpretability, which formalizes when counterfactual queries can be evaluated from a specific class of models.<n>We derive a complete graphical criterion that determines whether a given model architecture supports a given counterfactual query.
arXiv Detail & Related papers (2025-10-24T20:03:18Z) - A Logic of General Attention Using Edge-Conditioned Event Models (Extended Version) [1.6199400106794555]
We present the first general logic of attention.<n>Our work treats attention as a modality, like belief or awareness.<n>We illustrate our framework with examples of AI agents reasoning about human attentional biases.
arXiv Detail & Related papers (2025-05-20T15:56:34Z) - Counterfactual Explanations as Plans [6.445239204595516]
We look to provide a formal account of counterfactual explanations," based in terms of action sequences.
We then show that this naturally leads to an account of model reconciliation, which might take the form of the user correcting the agent's model, or suggesting actions to the agent's plan.
arXiv Detail & Related papers (2025-02-13T11:45:54Z) - States Hidden in Hidden States: LLMs Emerge Discrete State Representations Implicitly [72.24742240125369]
In this paper, we uncover the intrinsic ability to perform extended sequences of calculations without relying on chain-of-thought step-by-step solutions.
Remarkably, the most advanced models can directly output the results of two-digit number additions with lengths extending up to 15 addends.
arXiv Detail & Related papers (2024-07-16T06:27:22Z) - Diffusion Model with Cross Attention as an Inductive Bias for Disentanglement [58.9768112704998]
Disentangled representation learning strives to extract the intrinsic factors within observed data.
We introduce a new perspective and framework, demonstrating that diffusion models with cross-attention can serve as a powerful inductive bias.
This is the first work to reveal the potent disentanglement capability of diffusion models with cross-attention, requiring no complex designs.
arXiv Detail & Related papers (2024-02-15T05:07:54Z) - A Semantic Approach to Decidability in Epistemic Planning (Extended
Version) [72.77805489645604]
We use a novel semantic approach to achieve decidability.
Specifically, we augment the logic of knowledge S5$_n$ and with an interaction axiom called (knowledge) commutativity.
We prove that our framework admits a finitary non-fixpoint characterization of common knowledge, which is of independent interest.
arXiv Detail & Related papers (2023-07-28T11:26:26Z) - Suspected Object Matters: Rethinking Model's Prediction for One-stage
Visual Grounding [93.82542533426766]
We propose a Suspected Object Transformation mechanism (SOT) to encourage the target object selection among the suspected ones.
SOT can be seamlessly integrated into existing CNN and Transformer-based one-stage visual grounders.
Extensive experiments demonstrate the effectiveness of our proposed method.
arXiv Detail & Related papers (2022-03-10T06:41:07Z) - Modeling Event Plausibility with Consistent Conceptual Abstraction [29.69958315418181]
We show that Transformer-based plausibility models are markedly inconsistent across the conceptual classes of a lexical hierarchy.
We present a simple post-hoc method of forcing model consistency that improves correlation with human plausibility.
arXiv Detail & Related papers (2021-04-20T21:08:32Z) - Abstract Spatial-Temporal Reasoning via Probabilistic Abduction and
Execution [97.50813120600026]
Spatial-temporal reasoning is a challenging task in Artificial Intelligence (AI)
Recent works have focused on an abstract reasoning task of this kind -- Raven's Progressive Matrices ( RPM)
We propose a neuro-symbolic Probabilistic Abduction and Execution learner (PrAE) learner.
arXiv Detail & Related papers (2021-03-26T02:42:18Z) - SparseBERT: Rethinking the Importance Analysis in Self-attention [107.68072039537311]
Transformer-based models are popular for natural language processing (NLP) tasks due to its powerful capacity.
Attention map visualization of a pre-trained model is one direct method for understanding self-attention mechanism.
We propose a Differentiable Attention Mask (DAM) algorithm, which can be also applied in guidance of SparseBERT design.
arXiv Detail & Related papers (2021-02-25T14:13:44Z) - On the Dynamics of Training Attention Models [30.85940880569692]
We study the dynamics of training a simple attention-based classification model using gradient descent.
We prove that training must converge to attending to the discriminative words when the attention output is classified by a linear classifier.
arXiv Detail & Related papers (2020-11-19T18:55:30Z) - Attention or memory? Neurointerpretable agents in space and time [0.0]
We design a model incorporating a self-attention mechanism that implements task-state representations in semantic feature-space.
To evaluate the agent's selective properties, we add a large volume of task-irrelevant features to observations.
In line with neuroscience predictions, self-attention leads to increased robustness to noise compared to benchmark models.
arXiv Detail & Related papers (2020-07-09T15:04:26Z) - Towards Transparent and Explainable Attention Models [34.0557018891191]
We first explain why current attention mechanisms in LSTM based encoders can neither provide a faithful nor a plausible explanation of the model's predictions.
We propose a modified LSTM cell with a diversity-driven training objective that ensures that the hidden representations learned at different time steps are diverse.
Human evaluations indicate that the attention distributions learned by our model offer a plausible explanation of the model's predictions.
arXiv Detail & Related papers (2020-04-29T14:47:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.