Learning to cooperate: Emergent communication in multi-agent navigation
- URL: http://arxiv.org/abs/2004.01097v2
- Date: Tue, 30 Jun 2020 15:13:39 GMT
- Title: Learning to cooperate: Emergent communication in multi-agent navigation
- Authors: Ivana Kaji\'c, Eser Ayg\"un and Doina Precup
- Abstract summary: We show that agents performing a cooperative navigation task learn an interpretable communication protocol.
An analysis of the agents' policies reveals that emergent signals spatially cluster the state space.
Using populations of agents, we show that the emergent protocol has basic compositional structure.
- Score: 49.11609702016523
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Emergent communication in artificial agents has been studied to understand
language evolution, as well as to develop artificial systems that learn to
communicate with humans. We show that agents performing a cooperative
navigation task in various gridworld environments learn an interpretable
communication protocol that enables them to efficiently, and in many cases,
optimally, solve the task. An analysis of the agents' policies reveals that
emergent signals spatially cluster the state space, with signals referring to
specific locations and spatial directions such as "left", "up", or "upper left
room". Using populations of agents, we show that the emergent protocol has
basic compositional structure, thus exhibiting a core property of natural
language.
Related papers
- Bidirectional Emergent Language in Situated Environments [4.950411915351642]
We introduce two novel cooperative environments: Multi-Agent Pong and Collectors.
optimal performance requires the emergence of a communication protocol, but moderate success can be achieved without one.
We find that the emerging communication is sparse, with the agents only generating meaningful messages and acting upon incoming messages in states where they cannot succeed without coordination.
arXiv Detail & Related papers (2024-08-26T21:25:44Z) - Symbolic Learning Enables Self-Evolving Agents [55.625275970720374]
We introduce agent symbolic learning, a systematic framework that enables language agents to optimize themselves on their own.
Agent symbolic learning is designed to optimize the symbolic network within language agents by mimicking two fundamental algorithms in connectionist learning.
We conduct proof-of-concept experiments on both standard benchmarks and complex real-world tasks.
arXiv Detail & Related papers (2024-06-26T17:59:18Z) - Learning Multi-Agent Communication from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
Our proposed approach, CommFormer, efficiently optimize the communication graph and concurrently refines architectural parameters through gradient descent in an end-to-end manner.
arXiv Detail & Related papers (2024-05-14T12:40:25Z) - Building Cooperative Embodied Agents Modularly with Large Language
Models [104.57849816689559]
We address challenging multi-agent cooperation problems with decentralized control, raw sensory observations, costly communication, and multi-objective tasks instantiated in various embodied environments.
We harness the commonsense knowledge, reasoning ability, language comprehension, and text generation prowess of LLMs and seamlessly incorporate them into a cognitive-inspired modular framework.
Our experiments on C-WAH and TDW-MAT demonstrate that CoELA driven by GPT-4 can surpass strong planning-based methods and exhibit emergent effective communication.
arXiv Detail & Related papers (2023-07-05T17:59:27Z) - Cooperative Multi-Agent Learning for Navigation via Structured State
Abstraction [35.137788512033694]
This work proposes a novel neural network architecture for learning an adaptive state space abstraction and a communication protocol among agents participating in navigation tasks.
The goal is to come up with an adaptive abstractor that significantly reduces the size of the state space to be explored, without degradation in the policy performance.
arXiv Detail & Related papers (2023-06-20T07:06:17Z) - Learning to Ground Multi-Agent Communication with Autoencoders [43.22048280036316]
Communication requires a common language, a lingua franca, between agents.
We demonstrate a simple way to ground language in learned representations.
We find that a standard representation learning algorithm is sufficient for arriving at a grounded common language.
arXiv Detail & Related papers (2021-10-28T17:57:26Z) - Interpretation of Emergent Communication in Heterogeneous Collaborative
Embodied Agents [83.52684405389445]
We introduce the collaborative multi-object navigation task CoMON.
In this task, an oracle agent has detailed environment information in the form of a map.
It communicates with a navigator agent that perceives the environment visually and is tasked to find a sequence of goals.
We show that the emergent communication can be grounded to the agent observations and the spatial structure of the 3D environment.
arXiv Detail & Related papers (2021-10-12T06:56:11Z) - Few-shot Language Coordination by Modeling Theory of Mind [95.54446989205117]
We study the task of few-shot $textitlanguage coordination$.
We require the lead agent to coordinate with a $textitpopulation$ of agents with different linguistic abilities.
This requires the ability to model the partner's beliefs, a vital component of human communication.
arXiv Detail & Related papers (2021-07-12T19:26:11Z) - Networked Multi-Agent Reinforcement Learning with Emergent Communication [18.47483427884452]
Multi-Agent Reinforcement Learning (MARL) methods find optimal policies for agents that operate in the presence of other learning agents.
One way to coordinate is by learning to communicate with each other.
Can the agents develop a language while learning to perform a common task?
arXiv Detail & Related papers (2020-04-06T16:13:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.