Curriculum Learning for Cooperation in Multi-Agent Reinforcement
Learning
- URL: http://arxiv.org/abs/2312.11768v1
- Date: Tue, 19 Dec 2023 00:59:16 GMT
- Title: Curriculum Learning for Cooperation in Multi-Agent Reinforcement
Learning
- Authors: Rupali Bhati and Sai Krishna Gottipati and Clod\'eric Mars and Matthew
E. Taylor
- Abstract summary: In a competitive setting, a learning agent can be trained by making it compete with a curriculum of increasingly skilled opponents.
A general intelligent agent should also be able to learn to act around other agents and cooperate with them to achieve common goals.
In this paper, we aim to answer the question of what kind of cooperative teammate, and a curriculum of teammates should a learning agent be trained with to achieve these two objectives.
- Score: 7.336421550829047
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: While there has been significant progress in curriculum learning and
continuous learning for training agents to generalize across a wide variety of
environments in the context of single-agent reinforcement learning, it is
unclear if these algorithms would still be valid in a multi-agent setting. In a
competitive setting, a learning agent can be trained by making it compete with
a curriculum of increasingly skilled opponents. However, a general intelligent
agent should also be able to learn to act around other agents and cooperate
with them to achieve common goals. When cooperating with other agents, the
learning agent must (a) learn how to perform the task (or subtask), and (b)
increase the overall team reward. In this paper, we aim to answer the question
of what kind of cooperative teammate, and a curriculum of teammates should a
learning agent be trained with to achieve these two objectives. Our results on
the game Overcooked show that a pre-trained teammate who is less skilled is the
best teammate for overall team reward but the worst for the learning of the
agent. Moreover, somewhat surprisingly, a curriculum of teammates with
decreasing skill levels performs better than other types of curricula.
Related papers
- Multi-Agent Training for Pommerman: Curriculum Learning and Population-based Self-Play Approach [11.740631954398292]
Pommerman is an ideal benchmark for multi-agent training, providing a battleground for two teams with communication capabilities among allied agents.
This study introduces a system designed to train multi-agent systems to play Pommerman using a combination of curriculum learning and population-based self-play.
arXiv Detail & Related papers (2024-06-30T11:14:29Z) - ProAgent: Building Proactive Cooperative Agents with Large Language
Models [89.53040828210945]
ProAgent is a novel framework that harnesses large language models to create proactive agents.
ProAgent can analyze the present state, and infer the intentions of teammates from observations.
ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various coordination scenarios.
arXiv Detail & Related papers (2023-08-22T10:36:56Z) - Exploring the Benefits of Teams in Multiagent Learning [5.334505575267924]
We propose a new model of multiagent teams for reinforcement learning (RL) agents inspired by organizational psychology (OP)
We find that agents divided into teams develop cooperative pro-social policies despite incentives to not cooperate.
Agents are better able to coordinate and learn emergent roles within their teams and achieve higher rewards compared to when the interests of all agents are aligned.
arXiv Detail & Related papers (2022-05-04T21:14:03Z) - Learning to Transfer Role Assignment Across Team Sizes [48.43860606706273]
We propose a framework to learn role assignment and transfer across team sizes.
We demonstrate that re-using the role-based credit assignment structure can foster the learning process of larger reinforcement learning teams.
arXiv Detail & Related papers (2022-04-17T11:22:01Z) - Hidden Agenda: a Social Deduction Game with Diverse Learned Equilibria [57.74495091445414]
Social deduction games offer an avenue to study how individuals might learn to synthesize potentially unreliable information about others.
In this work, we present Hidden Agenda, a two-team social deduction game that provides a 2D environment for studying learning agents in scenarios of unknown team alignment.
Reinforcement learning agents trained in Hidden Agenda show that agents can learn a variety of behaviors, including partnering and voting without need for communication in natural language.
arXiv Detail & Related papers (2022-01-05T20:54:10Z) - Two-stage training algorithm for AI robot soccer [2.0757564643017092]
Two-stage heterogeneous centralized training is proposed to improve the learning performance of heterogeneous agents.
The proposed method is applied to 5 versus 5 AI robot soccer for validation.
arXiv Detail & Related papers (2021-04-13T04:24:13Z) - Natural Emergence of Heterogeneous Strategies in Artificially
Intelligent Competitive Teams [0.0]
We develop a competitive multi agent environment called FortAttack in which two teams compete against each other.
We observe a natural emergence of heterogeneous behavior amongst homogeneous agents when such behavior can lead to the team's success.
We propose ensemble training, in which we utilize the evolved opponent strategies to train a single policy for friendly agents.
arXiv Detail & Related papers (2020-07-06T22:35:56Z) - Learning to Incentivize Other Learning Agents [73.03133692589532]
We show how to equip RL agents with the ability to give rewards directly to other agents, using a learned incentive function.
Such agents significantly outperform standard RL and opponent-shaping agents in challenging general-sum Markov games.
Our work points toward more opportunities and challenges along the path to ensure the common good in a multi-agent future.
arXiv Detail & Related papers (2020-06-10T20:12:38Z) - Learning from Learners: Adapting Reinforcement Learning Agents to be
Competitive in a Card Game [71.24825724518847]
We present a study on how popular reinforcement learning algorithms can be adapted to learn and to play a real-world implementation of a competitive multiplayer card game.
We propose specific training and validation routines for the learning agents, in order to evaluate how the agents learn to be competitive and explain how they adapt to each others' playing style.
arXiv Detail & Related papers (2020-04-08T14:11:05Z) - On Emergent Communication in Competitive Multi-Agent Teams [116.95067289206919]
We investigate whether competition for performance from an external, similar agent team could act as a social influence.
Our results show that an external competitive influence leads to improved accuracy and generalization, as well as faster emergence of communicative languages.
arXiv Detail & Related papers (2020-03-04T01:14:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.