Toward Collaborative Reinforcement Learning Agents that Communicate
Through Text-Based Natural Language
- URL: http://arxiv.org/abs/2107.09356v2
- Date: Wed, 21 Jul 2021 13:28:31 GMT
- Title: Toward Collaborative Reinforcement Learning Agents that Communicate
Through Text-Based Natural Language
- Authors: Kevin Eloff, Herman A. Engelbrecht
- Abstract summary: This paper considers text-based natural language as a novel form of communication between agents trained with reinforcement learning.
Inspired by the game of Blind Leads, we propose an environment where one agent uses natural language instructions to guide another through a maze.
- Score: 4.289574109162585
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Communication between agents in collaborative multi-agent settings is in
general implicit or a direct data stream. This paper considers text-based
natural language as a novel form of communication between multiple agents
trained with reinforcement learning. This could be considered first steps
toward a truly autonomous communication without the need to define a limited
set of instructions, and natural collaboration between humans and robots.
Inspired by the game of Blind Leads, we propose an environment where one agent
uses natural language instructions to guide another through a maze. We test the
ability of reinforcement learning agents to effectively communicate through
discrete word-level symbols and show that the agents are able to sufficiently
communicate through natural language with a limited vocabulary. Although the
communication is not always perfect English, the agents are still able to
navigate the maze. We achieve a BLEU score of 0.85, which is an improvement of
0.61 over randomly generated sequences while maintaining a 100% maze completion
rate. This is a 3.5 times the performance of the random baseline using our
reference set.
Related papers
- Progressively Efficient Learning [58.6490456517954]
We develop a novel learning framework named Communication-Efficient Interactive Learning (CEIL)
CEIL leads to emergence of a human-like pattern where the learner and the teacher communicate efficiently by exchanging increasingly more abstract intentions.
Agents trained with CEIL quickly master new tasks, outperforming non-hierarchical and hierarchical imitation learning by up to 50% and 20% in absolute success rate.
arXiv Detail & Related papers (2023-10-13T07:52:04Z) - Building Cooperative Embodied Agents Modularly with Large Language
Models [104.57849816689559]
We address challenging multi-agent cooperation problems with decentralized control, raw sensory observations, costly communication, and multi-objective tasks instantiated in various embodied environments.
We harness the commonsense knowledge, reasoning ability, language comprehension, and text generation prowess of LLMs and seamlessly incorporate them into a cognitive-inspired modular framework.
Our experiments on C-WAH and TDW-MAT demonstrate that CoELA driven by GPT-4 can surpass strong planning-based methods and exhibit emergent effective communication.
arXiv Detail & Related papers (2023-07-05T17:59:27Z) - Transforming Human-Centered AI Collaboration: Redefining Embodied Agents
Capabilities through Interactive Grounded Language Instructions [23.318236094953072]
Human intelligence's adaptability is remarkable, allowing us to adjust to new tasks and multi-modal environments swiftly.
The research community is actively pursuing the development of interactive "embodied agents"
These agents must possess the ability to promptly request feedback in case communication breaks down or instructions are unclear.
arXiv Detail & Related papers (2023-05-18T07:51:33Z) - CAMEL: Communicative Agents for "Mind" Exploration of Large Language
Model Society [58.04479313658851]
This paper explores the potential of building scalable techniques to facilitate autonomous cooperation among communicative agents.
We propose a novel communicative agent framework named role-playing.
Our contributions include introducing a novel communicative agent framework, offering a scalable approach for studying the cooperative behaviors and capabilities of multi-agent systems.
arXiv Detail & Related papers (2023-03-31T01:09:00Z) - Linking Emergent and Natural Languages via Corpus Transfer [98.98724497178247]
We propose a novel way to establish a link by corpus transfer between emergent languages and natural languages.
Our approach showcases non-trivial transfer benefits for two different tasks -- language modeling and image captioning.
We also introduce a novel metric to predict the transferability of an emergent language by translating emergent messages to natural language captions grounded on the same images.
arXiv Detail & Related papers (2022-03-24T21:24:54Z) - Dynamic population-based meta-learning for multi-agent communication
with natural language [44.87604064505434]
We train agents that can coordinate with seen, unseen as well as human partners in a multi-agent communication environment involving natural language.
We propose a dynamic population-based meta-learning approach that builds such a population in an iterative manner.
We show that our agents outperform all prior work when communicating with seen partners and humans.
arXiv Detail & Related papers (2021-10-27T07:50:02Z) - Emergent Discrete Communication in Semantic Spaces [3.2280079436668996]
We propose a neural agent architecture that enables agents to communicate via discrete tokens derived from a learned, continuous space.
We show in a decision theoretic framework that our technique optimize communication over a wide range of scenarios, whereas one-hot tokens are only optimal under restrictive assumptions.
In self-play experiments, we validate that our trained agents learn to cluster tokens in semantically-meaningful ways, allowing them communicate in noisy environments.
arXiv Detail & Related papers (2021-08-04T03:32:48Z) - Few-shot Language Coordination by Modeling Theory of Mind [95.54446989205117]
We study the task of few-shot $textitlanguage coordination$.
We require the lead agent to coordinate with a $textitpopulation$ of agents with different linguistic abilities.
This requires the ability to model the partner's beliefs, a vital component of human communication.
arXiv Detail & Related papers (2021-07-12T19:26:11Z) - Self-play for Data Efficient Language Acquisition [20.86261546611472]
We exploit the symmetric nature of communication in order to improve the efficiency and quality of language acquisition in learning agents.
We show that using self-play as a substitute for direct supervision enables the agent to transfer its knowledge across roles.
arXiv Detail & Related papers (2020-10-10T02:09:19Z) - On the interaction between supervision and self-play in emergent
communication [82.290338507106]
We investigate the relationship between two categories of learning signals with the ultimate goal of improving sample efficiency.
We find that first training agents via supervised learning on human data followed by self-play outperforms the converse.
arXiv Detail & Related papers (2020-02-04T02:35:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.