Value-based CTDE Methods in Symmetric Two-team Markov Game: from
Cooperation to Team Competition
- URL: http://arxiv.org/abs/2211.11886v1
- Date: Mon, 21 Nov 2022 22:25:55 GMT
- Title: Value-based CTDE Methods in Symmetric Two-team Markov Game: from
Cooperation to Team Competition
- Authors: Pascal Leroy and Jonathan Pisane and Damien Ernst
- Abstract summary: We evaluate cooperative value-based methods in a mixed cooperative-competitive environment.
We selected three training methods based on the centralised training and decentralised execution paradigm.
For our experiments, we modified the StarCraft Multi-Agent Challenge environment to create competitive environments where both teams could learn and compete simultaneously.
- Score: 3.828689444527739
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we identify the best learning scenario to train a team of
agents to compete against multiple possible strategies of opposing teams. We
evaluate cooperative value-based methods in a mixed cooperative-competitive
environment. We restrict ourselves to the case of a symmetric, partially
observable, two-team Markov game. We selected three training methods based on
the centralised training and decentralised execution (CTDE) paradigm: QMIX,
MAVEN and QVMix. For each method, we considered three learning scenarios
differentiated by the variety of team policies encountered during training. For
our experiments, we modified the StarCraft Multi-Agent Challenge environment to
create competitive environments where both teams could learn and compete
simultaneously. Our results suggest that training against multiple evolving
strategies achieves the best results when, for scoring their performances,
teams are faced with several strategies.
Related papers
- CompetEvo: Towards Morphological Evolution from Competition [60.69068909395984]
We propose competitive evolution (CompetEvo), which co-evolves agents' designs and tactics in confrontation.
The results reveal that our method enables agents to evolve a more suitable design and strategy for fighting.
arXiv Detail & Related papers (2024-05-28T15:53:02Z) - All by Myself: Learning Individualized Competitive Behaviour with a
Contrastive Reinforcement Learning optimization [57.615269148301515]
In a competitive game scenario, a set of agents have to learn decisions that maximize their goals and minimize their adversaries' goals at the same time.
We propose a novel model composed of three neural layers that learn a representation of a competitive game, learn how to map the strategy of specific opponents, and how to disrupt them.
Our experiments demonstrate that our model achieves better performance when playing against offline, online, and competitive-specific models, in particular when playing against the same opponent multiple times.
arXiv Detail & Related papers (2023-10-02T08:11:07Z) - Benchmarking Robustness and Generalization in Multi-Agent Systems: A
Case Study on Neural MMO [50.58083807719749]
We present the results of the second Neural MMO challenge, hosted at IJCAI 2022, which received 1600+ submissions.
This competition targets robustness and generalization in multi-agent systems.
We will open-source our benchmark including the environment wrapper, baselines, a visualization tool, and selected policies for further research.
arXiv Detail & Related papers (2023-08-30T07:16:11Z) - Generating Teammates for Training Robust Ad Hoc Teamwork Agents via
Best-Response Diversity [6.940758395823777]
Ad hoc teamwork (AHT) is the challenge of designing a robust learner agent that effectively collaborates with unknown teammates.
Early approaches address the AHT challenge by training the learner with a diverse set of handcrafted teammate policies.
Recent approaches attempted to improve the robustness of the learner by training it with teammate policies generated by optimising information-theoretic diversity metrics.
arXiv Detail & Related papers (2022-07-28T14:58:32Z) - Retrospective on the 2021 BASALT Competition on Learning from Human
Feedback [92.37243979045817]
The goal of the competition was to promote research towards agents that use learning from human feedback (LfHF) techniques to solve open-world tasks.
Rather than mandating the use of LfHF techniques, we described four tasks in natural language to be accomplished in the video game Minecraft.
Teams developed a diverse range of LfHF algorithms across a variety of possible human feedback types.
arXiv Detail & Related papers (2022-04-14T17:24:54Z) - Conditional Imitation Learning for Multi-Agent Games [89.897635970366]
We study the problem of conditional multi-agent imitation learning, where we have access to joint trajectory demonstrations at training time.
We propose a novel approach to address the difficulties of scalability and data scarcity.
Our model learns a low-rank subspace over ego and partner agent strategies, then infers and adapts to a new partner strategy by interpolating in the subspace.
arXiv Detail & Related papers (2022-01-05T04:40:13Z) - Coach-Player Multi-Agent Reinforcement Learning for Dynamic Team
Composition [88.26752130107259]
In real-world multiagent systems, agents with different capabilities may join or leave without altering the team's overarching goals.
We propose COPA, a coach-player framework to tackle this problem.
We 1) adopt the attention mechanism for both the coach and the players; 2) propose a variational objective to regularize learning; and 3) design an adaptive communication method to let the coach decide when to communicate with the players.
arXiv Detail & Related papers (2021-05-18T17:27:37Z) - Learning to Play Soccer From Scratch: Sample-Efficient Emergent
Coordination through Curriculum-Learning and Competition [1.675857332621569]
This work proposes a scheme that allows learning complex multi-agent behaviors in a sample efficient manner, applied to 2v2 soccer.
The problem is formulated as a Markov game, and solved using deep reinforcement learning.
Our results show that high quality soccer play can be obtained with our approach in just under 40M interactions.
arXiv Detail & Related papers (2021-03-09T01:57:16Z) - Multi-Agent Coordination in Adversarial Environments through Signal
Mediated Strategies [37.00818384785628]
Team members can coordinate their strategies before the beginning of the game, but are unable to communicate during the playing phase of the game.
In this setting, model-free RL methods are oftentimes unable to capture coordination because agents' policies are executed in a decentralized fashion.
We show convergence to coordinated equilibria in cases where previous state-of-the-art multi-agent RL algorithms did not.
arXiv Detail & Related papers (2021-02-09T18:44:16Z) - An analysis of Reinforcement Learning applied to Coach task in IEEE Very
Small Size Soccer [2.5400028272658144]
This paper proposes an end-to-end approach for the coaching task based on Reinforcement Learning (RL)
We trained two RL policies against three different teams in a simulated environment.
Our results were assessed against one of the top teams of the VSSS league.
arXiv Detail & Related papers (2020-11-23T23:10:06Z) - Natural Emergence of Heterogeneous Strategies in Artificially
Intelligent Competitive Teams [0.0]
We develop a competitive multi agent environment called FortAttack in which two teams compete against each other.
We observe a natural emergence of heterogeneous behavior amongst homogeneous agents when such behavior can lead to the team's success.
We propose ensemble training, in which we utilize the evolved opponent strategies to train a single policy for friendly agents.
arXiv Detail & Related papers (2020-07-06T22:35:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.