Self-Explaining Deviations for Coordination
- URL: http://arxiv.org/abs/2207.12322v1
- Date: Wed, 13 Jul 2022 20:56:59 GMT
- Title: Self-Explaining Deviations for Coordination
- Authors: Hengyuan Hu, Samuel Sokota, David Wu, Anton Bakhtin, Andrei Lupu,
Brandon Cui, Jakob N. Foerster
- Abstract summary: We focus on a specific subclass of coordination problems in which humans are able to discover self-explaining deviations (SEDs)
SEDs are actions that deviate from the common understanding of what reasonable behavior would be in normal circumstances.
We introduce a novel algorithm, improvement maximizing self-explaining deviations (IMPROVISED), to perform SEDs.
- Score: 31.94421561348329
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Fully cooperative, partially observable multi-agent problems are ubiquitous
in the real world. In this paper, we focus on a specific subclass of
coordination problems in which humans are able to discover self-explaining
deviations (SEDs). SEDs are actions that deviate from the common understanding
of what reasonable behavior would be in normal circumstances. They are taken
with the intention of causing another agent or other agents to realize, using
theory of mind, that the circumstance must be abnormal. We first motivate SED
with a real world example and formalize its definition. Next, we introduce a
novel algorithm, improvement maximizing self-explaining deviations
(IMPROVISED), to perform SEDs. Lastly, we evaluate IMPROVISED both in an
illustrative toy setting and the popular benchmark setting Hanabi, where it is
the first method to produce so called finesse plays, which are regarded as one
of the more iconic examples of human theory of mind.
Related papers
- Action Inference by Maximising Evidence: Zero-Shot Imitation from
Observation with World Models [9.583751440005118]
We propose Action Inference by Maximising Evidence (AIME) to replicate this behaviour using world models.
AIME consists of two distinct phases. In the first phase, the agent learns a world model from its past experience to understand its own body by maximising the ELBO.
While in the second phase, the agent is given some observation-only demonstrations of an expert performing a novel task and tries to imitate the expert's behaviour.
Our method is "zero-shot" in the sense that it does not require further training for the world model or online interactions with the environment after given the demonstration
arXiv Detail & Related papers (2023-12-04T16:43:36Z) - UNcommonsense Reasoning: Abductive Reasoning about Uncommon Situations [62.71847873326847]
We investigate the ability to model unusual, unexpected, and unlikely situations.
Given a piece of context with an unexpected outcome, this task requires reasoning abductively to generate an explanation.
We release a new English language corpus called UNcommonsense.
arXiv Detail & Related papers (2023-11-14T19:00:55Z) - Habits and goals in synergy: a variational Bayesian framework for
behavior [22.461524318820672]
How to behave efficiently and flexibly is a central problem for understanding biological agents and creating intelligent embodied AI.
It has been well known that behavior can be classified as two types: reward-maximizing habitual behavior, which is fast while inflexible; and goal-directed behavior, which is flexible while slow.
We propose to bridge the gap between the two behaviors, drawing on the principles of variational Bayesian theory.
arXiv Detail & Related papers (2023-04-11T06:28:14Z) - Resonating Minds -- Emergent Collaboration Through Hierarchical Active
Inference [0.0]
We investigate how efficient, automatic coordination processes at the level of mental states (intentions, goals) can lead to collaborative situated problem-solving.
We present a model of hierarchical active inference for collaborative agents (HAICA)
We show that belief resonance and active inference allow for quick and efficient agent coordination, and thus can serve as a building block for collaborative cognitive agents.
arXiv Detail & Related papers (2021-12-02T13:23:44Z) - Nested Counterfactual Identification from Arbitrary Surrogate
Experiments [95.48089725859298]
We study the identification of nested counterfactuals from an arbitrary combination of observations and experiments.
Specifically, we prove the counterfactual unnesting theorem (CUT), which allows one to map arbitrary nested counterfactuals to unnested ones.
arXiv Detail & Related papers (2021-07-07T12:51:04Z) - Modeling the Mistakes of Boundedly Rational Agents Within a Bayesian
Theory of Mind [32.66203057545608]
We extend the Bayesian Theory of Mind framework to model boundedly rational agents who may have mistaken goals, plans, and actions.
We present experiments eliciting human goal inferences in two domains: (i) a gridworld puzzle with gems locked behind doors, and (ii) a block-stacking domain.
arXiv Detail & Related papers (2021-06-24T18:00:03Z) - End-to-End Learning and Intervention in Games [60.41921763076017]
We provide a unified framework for learning and intervention in games.
We propose two approaches, respectively based on explicit and implicit differentiation.
The analytical results are validated using several real-world problems.
arXiv Detail & Related papers (2020-10-26T18:39:32Z) - Machine Common Sense [77.34726150561087]
Machine common sense remains a broad, potentially unbounded problem in artificial intelligence (AI)
This article deals with the aspects of modeling commonsense reasoning focusing on such domain as interpersonal interactions.
arXiv Detail & Related papers (2020-06-15T13:59:47Z) - Agent-Based Simulation of Collective Cooperation: From Experiment to
Model [0.0]
We present an experiment to observe what happens when humans pass through a dense static crowd.
We derive a model that incorporates agents' perception and cognitive processing of a situation that needs cooperation.
Agents' ability to successfully get through a dense crowd emerges as an effect of the psychological model.
arXiv Detail & Related papers (2020-05-26T13:29:08Z) - Intrinsic Motivation for Encouraging Synergistic Behavior [55.10275467562764]
We study the role of intrinsic motivation as an exploration bias for reinforcement learning in sparse-reward synergistic tasks.
Our key idea is that a good guiding principle for intrinsic motivation in synergistic tasks is to take actions which affect the world in ways that would not be achieved if the agents were acting on their own.
arXiv Detail & Related papers (2020-02-12T19:34:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.