Few-shot Language Coordination by Modeling Theory of Mind
- URL: http://arxiv.org/abs/2107.05697v1
- Date: Mon, 12 Jul 2021 19:26:11 GMT
- Title: Few-shot Language Coordination by Modeling Theory of Mind
- Authors: Hao Zhu, Graham Neubig, Yonatan Bisk
- Abstract summary: We study the task of few-shot $textitlanguage coordination$.
We require the lead agent to coordinate with a $textitpopulation$ of agents with different linguistic abilities.
This requires the ability to model the partner's beliefs, a vital component of human communication.
- Score: 95.54446989205117
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: $\textit{No man is an island.}$ Humans communicate with a large community by
coordinating with different interlocutors within short conversations. This
ability has been understudied by the research on building neural communicative
agents. We study the task of few-shot $\textit{language coordination}$: agents
quickly adapting to their conversational partners' language abilities.
Different from current communicative agents trained with self-play, we require
the lead agent to coordinate with a $\textit{population}$ of agents with
different linguistic abilities, quickly adapting to communicate with unseen
agents in the population. This requires the ability to model the partner's
beliefs, a vital component of human communication. Drawing inspiration from
theory-of-mind (ToM; Premack& Woodruff (1978)), we study the effect of the
speaker explicitly modeling the listeners' mental states. The speakers, as
shown in our experiments, acquire the ability to predict the reactions of their
partner, which helps it generate instructions that concisely express its
communicative goal. We examine our hypothesis that the instructions generated
with ToM modeling yield better communication performance in both a referential
game and a language navigation task. Positive results from our experiments hint
at the importance of explicitly modeling communication as a socio-pragmatic
progress.
Related papers
- Interactive Dialogue Agents via Reinforcement Learning on Hindsight Regenerations [58.65755268815283]
Many real dialogues are interactive, meaning an agent's utterances will influence their conversational partner, elicit information, or change their opinion.
We use this fact to rewrite and augment existing suboptimal data, and train via offline reinforcement learning (RL) an agent that outperforms both prompting and learning from unaltered human demonstrations.
Our results in a user study with real humans show that our approach greatly outperforms existing state-of-the-art dialogue agents.
arXiv Detail & Related papers (2024-11-07T21:37:51Z) - GOMA: Proactive Embodied Cooperative Communication via Goal-Oriented Mental Alignment [72.96949760114575]
We propose a novel cooperative communication framework, Goal-Oriented Mental Alignment (GOMA)
GOMA formulates verbal communication as a planning problem that minimizes the misalignment between parts of agents' mental states that are relevant to the goals.
We evaluate our approach against strong baselines in two challenging environments, Overcooked (a multiplayer game) and VirtualHome (a household simulator)
arXiv Detail & Related papers (2024-03-17T03:52:52Z) - CAMEL: Communicative Agents for "Mind" Exploration of Large Language
Model Society [58.04479313658851]
This paper explores the potential of building scalable techniques to facilitate autonomous cooperation among communicative agents.
We propose a novel communicative agent framework named role-playing.
Our contributions include introducing a novel communicative agent framework, offering a scalable approach for studying the cooperative behaviors and capabilities of multi-agent systems.
arXiv Detail & Related papers (2023-03-31T01:09:00Z) - Computational Language Acquisition with Theory of Mind [84.2267302901888]
We build language-learning agents equipped with Theory of Mind (ToM) and measure its effects on the learning process.
We find that training speakers with a highly weighted ToM listener component leads to performance gains in our image referential game setting.
arXiv Detail & Related papers (2023-03-02T18:59:46Z) - Communication Drives the Emergence of Language Universals in Neural
Agents: Evidence from the Word-order/Case-marking Trade-off [3.631024220680066]
We propose a new Neural-agent Language Learning and Communication framework (NeLLCom) where pairs of speaking and listening agents first learn a miniature language.
We succeed in replicating the trade-off with the new framework without hard-coding specific biases in the agents.
arXiv Detail & Related papers (2023-01-30T17:22:33Z) - Interpretation of Emergent Communication in Heterogeneous Collaborative
Embodied Agents [83.52684405389445]
We introduce the collaborative multi-object navigation task CoMON.
In this task, an oracle agent has detailed environment information in the form of a map.
It communicates with a navigator agent that perceives the environment visually and is tasked to find a sequence of goals.
We show that the emergent communication can be grounded to the agent observations and the spatial structure of the 3D environment.
arXiv Detail & Related papers (2021-10-12T06:56:11Z) - Self-play for Data Efficient Language Acquisition [20.86261546611472]
We exploit the symmetric nature of communication in order to improve the efficiency and quality of language acquisition in learning agents.
We show that using self-play as a substitute for direct supervision enables the agent to transfer its knowledge across roles.
arXiv Detail & Related papers (2020-10-10T02:09:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.