Two-stage training algorithm for AI robot soccer
- URL: http://arxiv.org/abs/2104.05931v1
- Date: Tue, 13 Apr 2021 04:24:13 GMT
- Title: Two-stage training algorithm for AI robot soccer
- Authors: Taeyoung Kim, Luiz Felipe Vecchietti, Kyujin Choi, Sanem Sariel,
Dongsoo Har
- Abstract summary: Two-stage heterogeneous centralized training is proposed to improve the learning performance of heterogeneous agents.
The proposed method is applied to 5 versus 5 AI robot soccer for validation.
- Score: 2.0757564643017092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In multi-agent reinforcement learning, the cooperative learning behavior of
agents is very important. In the field of heterogeneous multi-agent
reinforcement learning, cooperative behavior among different types of agents in
a group is pursued. Learning a joint-action set during centralized training is
an attractive way to obtain such cooperative behavior, however, this method
brings limited learning performance with heterogeneous agents. To improve the
learning performance of heterogeneous agents during centralized training,
two-stage heterogeneous centralized training which allows the training of
multiple roles of heterogeneous agents is proposed. During training, two
training processes are conducted in a series. One of the two stages is to
attempt training each agent according to its role, aiming at the maximization
of individual role rewards. The other is for training the agents as a whole to
make them learn cooperative behaviors while attempting to maximize shared
collective rewards, e.g., team rewards. Because these two training processes
are conducted in a series in every timestep, agents can learn how to maximize
role rewards and team rewards simultaneously. The proposed method is applied to
5 versus 5 AI robot soccer for validation. Simulation results show that the
proposed method can train the robots of the robot soccer team effectively,
achieving higher role rewards and higher team rewards as compared to other
approaches that can be used to solve problems of training cooperative
multi-agent.
Related papers
- Multi-Agent Training for Pommerman: Curriculum Learning and Population-based Self-Play Approach [11.740631954398292]
Pommerman is an ideal benchmark for multi-agent training, providing a battleground for two teams with communication capabilities among allied agents.
This study introduces a system designed to train multi-agent systems to play Pommerman using a combination of curriculum learning and population-based self-play.
arXiv Detail & Related papers (2024-06-30T11:14:29Z) - ProAgent: Building Proactive Cooperative Agents with Large Language
Models [89.53040828210945]
ProAgent is a novel framework that harnesses large language models to create proactive agents.
ProAgent can analyze the present state, and infer the intentions of teammates from observations.
ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various coordination scenarios.
arXiv Detail & Related papers (2023-08-22T10:36:56Z) - AgentVerse: Facilitating Multi-Agent Collaboration and Exploring
Emergent Behaviors [93.38830440346783]
We propose a multi-agent framework framework that can collaboratively adjust its composition as a greater-than-the-sum-of-its-parts system.
Our experiments demonstrate that framework framework can effectively deploy multi-agent groups that outperform a single agent.
In view of these behaviors, we discuss some possible strategies to leverage positive ones and mitigate negative ones for improving the collaborative potential of multi-agent groups.
arXiv Detail & Related papers (2023-08-21T16:47:11Z) - Learning Heterogeneous Agent Cooperation via Multiagent League Training [6.801749815385998]
This work proposes a general-purpose reinforcement learning algorithm named Heterogeneous League Training (HLT) to address heterogeneous multiagent problems.
HLT keeps track of a pool of policies that agents have explored during training, gathering a league of heterogeneous policies to facilitate future policy optimization.
A hyper-network is introduced to increase the diversity of agent behaviors when collaborating with teammates having different levels of cooperation skills.
arXiv Detail & Related papers (2022-11-13T13:57:15Z) - ELIGN: Expectation Alignment as a Multi-Agent Intrinsic Reward [29.737986509769808]
We propose a self-supervised intrinsic reward ELIGN - expectation alignment.
Similar to how animals collaborate in a decentralized manner with those in their vicinity, agents trained with expectation alignment learn behaviors that match their neighbors' expectations.
We show that agent coordination improves through expectation alignment because agents learn to divide tasks amongst themselves, break coordination symmetries, and confuse adversaries.
arXiv Detail & Related papers (2022-10-09T22:24:44Z) - LDSA: Learning Dynamic Subtask Assignment in Cooperative Multi-Agent
Reinforcement Learning [122.47938710284784]
We propose a novel framework for learning dynamic subtask assignment (LDSA) in cooperative MARL.
To reasonably assign agents to different subtasks, we propose an ability-based subtask selection strategy.
We show that LDSA learns reasonable and effective subtask assignment for better collaboration.
arXiv Detail & Related papers (2022-05-05T10:46:16Z) - Learning to Transfer Role Assignment Across Team Sizes [48.43860606706273]
We propose a framework to learn role assignment and transfer across team sizes.
We demonstrate that re-using the role-based credit assignment structure can foster the learning process of larger reinforcement learning teams.
arXiv Detail & Related papers (2022-04-17T11:22:01Z) - Coach-assisted Multi-Agent Reinforcement Learning Framework for
Unexpected Crashed Agents [120.91291581594773]
We present a formal formulation of a cooperative multi-agent reinforcement learning system with unexpected crashes.
We propose a coach-assisted multi-agent reinforcement learning framework, which introduces a virtual coach agent to adjust the crash rate during training.
To the best of our knowledge, this work is the first to study the unexpected crashes in the multi-agent system.
arXiv Detail & Related papers (2022-03-16T08:22:45Z) - Evaluating Generalization and Transfer Capacity of Multi-Agent
Reinforcement Learning Across Variable Number of Agents [0.0]
Multi-agent Reinforcement Learning (MARL) problems often require cooperation among agents in order to solve a task.
Centralization and decentralization are two approaches used for cooperation in MARL.
We adopt centralized training with decentralized execution paradigm and investigate the generalization and transfer capacity of the trained models across variable number of agents.
arXiv Detail & Related papers (2021-11-28T15:29:46Z) - UneVEn: Universal Value Exploration for Multi-Agent Reinforcement
Learning [53.73686229912562]
We propose a novel MARL approach called Universal Value Exploration (UneVEn)
UneVEn learns a set of related tasks simultaneously with a linear decomposition of universal successor features.
Empirical results on a set of exploration games, challenging cooperative predator-prey tasks requiring significant coordination among agents, and StarCraft II micromanagement benchmarks show that UneVEn can solve tasks where other state-of-the-art MARL methods fail.
arXiv Detail & Related papers (2020-10-06T19:08:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.