Learning Translations: Emergent Communication Pretraining for
Cooperative Language Acquisition
- URL: http://arxiv.org/abs/2402.16247v1
- Date: Mon, 26 Feb 2024 02:13:36 GMT
- Title: Learning Translations: Emergent Communication Pretraining for
Cooperative Language Acquisition
- Authors: Dylan Cope and Peter McBurney
- Abstract summary: In Emergent Communication (EC) agents learn to communicate with one another, but the protocols that they develop are specialised to their training community.
This observation led to research into Zero-Shot Coordination (ZSC) for learning communication strategies that are robust to agents not encountered during training.
We propose a novel AI challenge called a Cooperative Language Acquisition Problem (CLAP) in which the ZSC assumptions are relaxed by allowing a 'joiner' agent to learn from a dataset of interactions between agents in a target community.
- Score: 0.7832189413179361
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In Emergent Communication (EC) agents learn to communicate with one another,
but the protocols that they develop are specialised to their training
community. This observation led to research into Zero-Shot Coordination (ZSC)
for learning communication strategies that are robust to agents not encountered
during training. However, ZSC typically assumes that no prior data is available
about the agents that will be encountered in the zero-shot setting. In many
cases, this presents an unnecessarily hard problem and rules out communication
via preestablished conventions. We propose a novel AI challenge called a
Cooperative Language Acquisition Problem (CLAP) in which the ZSC assumptions
are relaxed by allowing a 'joiner' agent to learn from a dataset of
interactions between agents in a target community. We propose and compare two
methods for solving CLAPs: Imitation Learning (IL), and Emergent Communication
pretraining and Translation Learning (ECTL), in which an agent is trained in
self-play with EC and then learns from the data to translate between the
emergent protocol and the target community's protocol.
Related papers
- Batch Selection and Communication for Active Learning with Edge Labeling [54.64985724916654]
Communication-Constrained Bayesian Active Knowledge Distillation (CC-BAKD)
This work introduces Communication-Constrained Bayesian Active Knowledge Distillation (CC-BAKD)
arXiv Detail & Related papers (2023-11-14T10:23:00Z) - AC2C: Adaptively Controlled Two-Hop Communication for Multi-Agent
Reinforcement Learning [4.884877440051105]
We propose a novel communication protocol called Adaptively Controlled Two-Hop Communication (AC2C)
AC2C employs an adaptive two-hop communication strategy to enable long-range information exchange among agents to boost performance.
We evaluate AC2C on three cooperative multi-agent tasks, and the experimental results show that it outperforms relevant baselines with lower communication costs.
arXiv Detail & Related papers (2023-02-24T09:00:34Z) - Over-communicate no more: Situated RL agents learn concise communication
protocols [78.28898217947467]
It is unclear how to design artificial agents that can learn to effectively and efficiently communicate with each other.
Much research on communication emergence uses reinforcement learning (RL)
We explore situated communication in a multi-step task, where the acting agent has to forgo an environmental action to communicate.
We find that while all tested pressures can disincentivise over-communication, situated communication does it most effectively and, unlike the cost on effort, does not negatively impact emergence.
arXiv Detail & Related papers (2022-11-02T21:08:14Z) - Coordinating Policies Among Multiple Agents via an Intelligent
Communication Channel [81.39444892747512]
In Multi-Agent Reinforcement Learning (MARL), specialized channels are often introduced that allow agents to communicate directly with one another.
We propose an alternative approach whereby agents communicate through an intelligent facilitator that learns to sift through and interpret signals provided by all agents to improve the agents' collective performance.
arXiv Detail & Related papers (2022-05-21T14:11:33Z) - Quasi-Equivalence Discovery for Zero-Shot Emergent Communication [63.175848843466845]
We present a novel problem setting and the Quasi-Equivalence Discovery algorithm that allows for zero-shot coordination (ZSC)
We show that these two factors lead to unique optimal ZSC policies in referential games.
QED can iteratively discover the symmetries in this setting and converges to the optimal ZSC policy.
arXiv Detail & Related papers (2021-03-14T23:42:37Z) - Exploring Zero-Shot Emergent Communication in Embodied Multi-Agent
Populations [59.608216900601384]
We study agents that learn to communicate via actuating their joints in a 3D environment.
We show that under realistic assumptions, a non-uniform distribution of intents and a common-knowledge energy cost, these agents can find protocols that generalize to novel partners.
arXiv Detail & Related papers (2020-10-29T19:23:10Z) - Learning to Communicate Using Counterfactual Reasoning [2.8110705488739676]
This paper introduces the novel multi-agent counterfactual communication learning (MACC) method.
MACC adapts counterfactual reasoning in order to overcome the credit assignment problem for communicating agents.
Our experiments show that MACC is able to outperform the state-of-the-art baselines in four different scenarios in the Particle environment.
arXiv Detail & Related papers (2020-06-12T14:02:04Z) - Learning Individually Inferred Communication for Multi-Agent Cooperation [37.56115000150748]
We propose Individually Inferred Communication (I2C) to enable agents to learn a prior for agent-agent communication.
The prior knowledge is learned via causal inference and realized by a feed-forward neural network.
I2C can not only reduce communication overhead but also improve the performance in a variety of multi-agent cooperative scenarios.
arXiv Detail & Related papers (2020-06-11T14:07:57Z) - Learning to cooperate: Emergent communication in multi-agent navigation [49.11609702016523]
We show that agents performing a cooperative navigation task learn an interpretable communication protocol.
An analysis of the agents' policies reveals that emergent signals spatially cluster the state space.
Using populations of agents, we show that the emergent protocol has basic compositional structure.
arXiv Detail & Related papers (2020-04-02T16:03:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.