Inferring the Goals of Communicating Agents from Actions and
Instructions
- URL: http://arxiv.org/abs/2306.16207v1
- Date: Wed, 28 Jun 2023 13:43:46 GMT
- Title: Inferring the Goals of Communicating Agents from Actions and
Instructions
- Authors: Lance Ying, Tan Zhi-Xuan, Vikash Mansinghka, Joshua B. Tenenbaum
- Abstract summary: We introduce a model of a cooperative team where one agent, the principal, may communicate natural language instructions about their shared plan to another agent, the assistant.
We show how a third person observer can infer the team's goal via multi-modal inverse planning from actions and instructions.
We evaluate this approach by comparing it with human goal inferences in a multi-agent gridworld, finding that our model's inferences closely correlate with human judgments.
- Score: 47.5816320484482
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When humans cooperate, they frequently coordinate their activity through both
verbal communication and non-verbal actions, using this information to infer a
shared goal and plan. How can we model this inferential ability? In this paper,
we introduce a model of a cooperative team where one agent, the principal, may
communicate natural language instructions about their shared plan to another
agent, the assistant, using GPT-3 as a likelihood function for instruction
utterances. We then show how a third person observer can infer the team's goal
via multi-modal Bayesian inverse planning from actions and instructions,
computing the posterior distribution over goals under the assumption that
agents will act and communicate rationally to achieve them. We evaluate this
approach by comparing it with human goal inferences in a multi-agent gridworld,
finding that our model's inferences closely correlate with human judgments (R =
0.96). When compared to inference from actions alone, we also find that
instructions lead to more rapid and less uncertain goal inference, highlighting
the importance of verbal communication for cooperative agents.
Related papers
- Pragmatic Instruction Following and Goal Assistance via Cooperative
Language-Guided Inverse Planning [52.91457780361305]
This paper introduces cooperative language-guided inverse plan search (CLIPS)
Our agent assists a human by modeling them as a cooperative planner who communicates joint plans to the assistant.
We evaluate these capabilities in two cooperative planning domains (Doors, Keys & Gems and VirtualHome)
arXiv Detail & Related papers (2024-02-27T23:06:53Z) - Control as Probabilistic Inference as an Emergent Communication
Mechanism in Multi-Agent Reinforcement Learning [7.0471949371778795]
This paper proposes a generative probabilistic model integrating emergent communication and reinforcement learning.
We show that the proposed PGM can infer meaningful messages to achieve the cooperative task.
arXiv Detail & Related papers (2023-07-11T03:53:46Z) - Building Cooperative Embodied Agents Modularly with Large Language
Models [104.57849816689559]
We address challenging multi-agent cooperation problems with decentralized control, raw sensory observations, costly communication, and multi-objective tasks instantiated in various embodied environments.
We harness the commonsense knowledge, reasoning ability, language comprehension, and text generation prowess of LLMs and seamlessly incorporate them into a cognitive-inspired modular framework.
Our experiments on C-WAH and TDW-MAT demonstrate that CoELA driven by GPT-4 can surpass strong planning-based methods and exhibit emergent effective communication.
arXiv Detail & Related papers (2023-07-05T17:59:27Z) - CAMEL: Communicative Agents for "Mind" Exploration of Large Language
Model Society [58.04479313658851]
This paper explores the potential of building scalable techniques to facilitate autonomous cooperation among communicative agents.
We propose a novel communicative agent framework named role-playing.
Our contributions include introducing a novel communicative agent framework, offering a scalable approach for studying the cooperative behaviors and capabilities of multi-agent systems.
arXiv Detail & Related papers (2023-03-31T01:09:00Z) - ToM2C: Target-oriented Multi-agent Communication and Cooperation with
Theory of Mind [18.85252946546942]
Theory of Mind (ToM) builds socially intelligent agents who are able to communicate and cooperate effectively.
We demonstrate the idea in two typical target-oriented multi-agent tasks: cooperative navigation and multi-sensor target coverage.
arXiv Detail & Related papers (2021-10-15T18:29:55Z) - Interpretation of Emergent Communication in Heterogeneous Collaborative
Embodied Agents [83.52684405389445]
We introduce the collaborative multi-object navigation task CoMON.
In this task, an oracle agent has detailed environment information in the form of a map.
It communicates with a navigator agent that perceives the environment visually and is tasked to find a sequence of goals.
We show that the emergent communication can be grounded to the agent observations and the spatial structure of the 3D environment.
arXiv Detail & Related papers (2021-10-12T06:56:11Z) - Few-shot Language Coordination by Modeling Theory of Mind [95.54446989205117]
We study the task of few-shot $textitlanguage coordination$.
We require the lead agent to coordinate with a $textitpopulation$ of agents with different linguistic abilities.
This requires the ability to model the partner's beliefs, a vital component of human communication.
arXiv Detail & Related papers (2021-07-12T19:26:11Z) - The Emergence of Adversarial Communication in Multi-Agent Reinforcement
Learning [6.18778092044887]
Many real-world problems require the coordination of multiple autonomous agents.
Recent work has shown the promise of Graph Neural Networks (GNNs) to learn explicit communication strategies that enable complex multi-agent coordination.
We show how a single self-interested agent is capable of learning highly manipulative communication strategies that allows it to significantly outperform a cooperative team of agents.
arXiv Detail & Related papers (2020-08-06T12:48:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.