Learning Heterogeneous Agent Cooperation via Multiagent League Training
- URL: http://arxiv.org/abs/2211.11616v2
- Date: Sun, 28 May 2023 15:38:03 GMT
- Title: Learning Heterogeneous Agent Cooperation via Multiagent League Training
- Authors: Qingxu Fu, Xiaolin Ai, Jianqiang Yi, Tenghai Qiu, Wanmai Yuan,
Zhiqiang Pu
- Abstract summary: This work proposes a general-purpose reinforcement learning algorithm named Heterogeneous League Training (HLT) to address heterogeneous multiagent problems.
HLT keeps track of a pool of policies that agents have explored during training, gathering a league of heterogeneous policies to facilitate future policy optimization.
A hyper-network is introduced to increase the diversity of agent behaviors when collaborating with teammates having different levels of cooperation skills.
- Score: 6.801749815385998
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many multiagent systems in the real world include multiple types of agents
with different abilities and functionality. Such heterogeneous multiagent
systems have significant practical advantages. However, they also come with
challenges compared with homogeneous systems for multiagent reinforcement
learning, such as the non-stationary problem and the policy version iteration
issue. This work proposes a general-purpose reinforcement learning algorithm
named Heterogeneous League Training (HLT) to address heterogeneous multiagent
problems. HLT keeps track of a pool of policies that agents have explored
during training, gathering a league of heterogeneous policies to facilitate
future policy optimization. Moreover, a hyper-network is introduced to increase
the diversity of agent behaviors when collaborating with teammates having
different levels of cooperation skills. We use heterogeneous benchmark tasks to
demonstrate that (1) HLT promotes the success rate in cooperative heterogeneous
tasks; (2) HLT is an effective approach to solving the policy version iteration
problem; (3) HLT provides a practical way to assess the difficulty of learning
each role in a heterogeneous team.
Related papers
- QTypeMix: Enhancing Multi-Agent Cooperative Strategies through Heterogeneous and Homogeneous Value Decomposition [11.170571181947274]
We propose QTypeMix, which divides the value decomposition process into homogeneous and heterogeneous stages.
The results of testing the proposed method on 14 maps from SMAC and SMACv2 show that QTypeMix achieves state-of-the-art performance in tasks of varying difficulty.
arXiv Detail & Related papers (2024-08-12T12:27:58Z) - Prioritized League Reinforcement Learning for Large-Scale Heterogeneous Multiagent Systems [11.017749510087059]
We propose a prioritized Heterogeneous League Reinforcement Learning (PHLRL) method to address large-scale heterogeneous cooperation problems.
We use Unreal Engine to design a large-scale heterogeneous cooperation benchmark named Large-Scale Multiagent Operation (LSMO)
arXiv Detail & Related papers (2024-03-26T19:21:50Z) - ProAgent: Building Proactive Cooperative Agents with Large Language
Models [89.53040828210945]
ProAgent is a novel framework that harnesses large language models to create proactive agents.
ProAgent can analyze the present state, and infer the intentions of teammates from observations.
ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various coordination scenarios.
arXiv Detail & Related papers (2023-08-22T10:36:56Z) - Heterogeneous Embodied Multi-Agent Collaboration [21.364827833498254]
Heterogeneous multi-agent tasks are common in real-world scenarios.
We propose the heterogeneous multi-agent tidying-up task, in which multiple heterogeneous agents collaborate to detect misplaced objects and place them in reasonable locations.
We propose the hierarchical decision model based on misplaced object detection, reasonable receptacle prediction, as well as the handshake-based group communication mechanism.
arXiv Detail & Related papers (2023-07-26T04:33:05Z) - Learning Reward Machines in Cooperative Multi-Agent Tasks [75.79805204646428]
This paper presents a novel approach to Multi-Agent Reinforcement Learning (MARL)
It combines cooperative task decomposition with the learning of reward machines (RMs) encoding the structure of the sub-tasks.
The proposed method helps deal with the non-Markovian nature of the rewards in partially observable environments.
arXiv Detail & Related papers (2023-03-24T15:12:28Z) - Heterogeneous Multi-agent Zero-Shot Coordination by Coevolution [41.23036865145942]
We study the heterogeneous zero-shot coordination (ZSC) problem for the first time.
We propose a general method based on coevolution, which coevolves two populations of agents and partners through three sub-processes: pairing, updating and selection.
arXiv Detail & Related papers (2022-08-09T16:16:28Z) - Hierarchical Reinforcement Learning with Opponent Modeling for
Distributed Multi-agent Cooperation [13.670618752160594]
Deep reinforcement learning (DRL) provides a promising approach for multi-agent cooperation through the interaction of the agents and environments.
Traditional DRL solutions suffer from the high dimensions of multiple agents with continuous action space during policy search.
We propose a hierarchical reinforcement learning approach with high-level decision-making and low-level individual control for efficient policy search.
arXiv Detail & Related papers (2022-06-25T19:09:29Z) - Policy Diagnosis via Measuring Role Diversity in Cooperative Multi-agent
RL [107.58821842920393]
We quantify the agent's behavior difference and build its relationship with the policy performance via bf Role Diversity
We find that the error bound in MARL can be decomposed into three parts that have a strong relation to the role diversity.
The decomposed factors can significantly impact policy optimization on three popular directions.
arXiv Detail & Related papers (2022-06-01T04:58:52Z) - LDSA: Learning Dynamic Subtask Assignment in Cooperative Multi-Agent
Reinforcement Learning [122.47938710284784]
We propose a novel framework for learning dynamic subtask assignment (LDSA) in cooperative MARL.
To reasonably assign agents to different subtasks, we propose an ability-based subtask selection strategy.
We show that LDSA learns reasonable and effective subtask assignment for better collaboration.
arXiv Detail & Related papers (2022-05-05T10:46:16Z) - Multi-agent Policy Optimization with Approximatively Synchronous
Advantage Estimation [55.96893934962757]
In multi-agent system, polices of different agents need to be evaluated jointly.
In current methods, value functions or advantage functions use counter-factual joint actions which are evaluated asynchronously.
In this work, we propose the approximatively synchronous advantage estimation.
arXiv Detail & Related papers (2020-12-07T07:29:19Z) - UneVEn: Universal Value Exploration for Multi-Agent Reinforcement
Learning [53.73686229912562]
We propose a novel MARL approach called Universal Value Exploration (UneVEn)
UneVEn learns a set of related tasks simultaneously with a linear decomposition of universal successor features.
Empirical results on a set of exploration games, challenging cooperative predator-prey tasks requiring significant coordination among agents, and StarCraft II micromanagement benchmarks show that UneVEn can solve tasks where other state-of-the-art MARL methods fail.
arXiv Detail & Related papers (2020-10-06T19:08:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.