Cooperation and Competition: Flocking with Evolutionary Multi-Agent
Reinforcement Learning
- URL: http://arxiv.org/abs/2209.04696v2
- Date: Tue, 13 Sep 2022 06:56:54 GMT
- Title: Cooperation and Competition: Flocking with Evolutionary Multi-Agent
Reinforcement Learning
- Authors: Yunxiao Guo, Xinjia Xie, Runhao Zhao, Chenglan Zhu, Jiangting Yin, Han
Long
- Abstract summary: We propose Evolutionary Multi-Agent Reinforcement Learning (EMARL) in flocking tasks.
EMARL combines cooperation and competition with little prior knowledge.
We show that EMARL significantly outperforms the full competition or cooperation methods.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Flocking is a very challenging problem in a multi-agent system; traditional
flocking methods also require complete knowledge of the environment and a
precise model for control. In this paper, we propose Evolutionary Multi-Agent
Reinforcement Learning (EMARL) in flocking tasks, a hybrid algorithm that
combines cooperation and competition with little prior knowledge. As for
cooperation, we design the agents' reward for flocking tasks according to the
boids model. While for competition, agents with high fitness are designed as
senior agents, and those with low fitness are designed as junior, letting
junior agents inherit the parameters of senior agents stochastically. To
intensify competition, we also design an evolutionary selection mechanism that
shows effectiveness on credit assignment in flocking tasks. Experimental
results in a range of challenging and self-contrast benchmarks demonstrate that
EMARL significantly outperforms the full competition or cooperation methods.
Related papers
- Scaling Large-Language-Model-based Multi-Agent Collaboration [75.5241464256688]
Pioneering advancements in large language model-powered agents have underscored the design pattern of multi-agent collaboration.
Inspired by the neural scaling law, this study investigates whether a similar principle applies to increasing agents in multi-agent collaboration.
arXiv Detail & Related papers (2024-06-11T11:02:04Z) - CompetEvo: Towards Morphological Evolution from Competition [60.69068909395984]
We propose competitive evolution (CompetEvo), which co-evolves agents' designs and tactics in confrontation.
The results reveal that our method enables agents to evolve a more suitable design and strategy for fighting.
arXiv Detail & Related papers (2024-05-28T15:53:02Z) - Joint Intrinsic Motivation for Coordinated Exploration in Multi-Agent
Deep Reinforcement Learning [0.0]
We propose an approach for rewarding strategies where agents collectively exhibit novel behaviors.
Jim rewards joint trajectories based on a centralized measure of novelty designed to function in continuous environments.
Results show that joint exploration is crucial for solving tasks where the optimal strategy requires a high level of coordination.
arXiv Detail & Related papers (2024-02-06T13:02:00Z) - CompeteAI: Understanding the Competition Dynamics in Large Language Model-based Agents [43.46476421809271]
Large language models (LLMs) have been widely used as agents to complete different tasks.
We propose a general framework for studying the competition between agents.
We then implement a practical competitive environment using GPT-4 to simulate a virtual town.
arXiv Detail & Related papers (2023-10-26T16:06:20Z) - Benchmarking Robustness and Generalization in Multi-Agent Systems: A
Case Study on Neural MMO [50.58083807719749]
We present the results of the second Neural MMO challenge, hosted at IJCAI 2022, which received 1600+ submissions.
This competition targets robustness and generalization in multi-agent systems.
We will open-source our benchmark including the environment wrapper, baselines, a visualization tool, and selected policies for further research.
arXiv Detail & Related papers (2023-08-30T07:16:11Z) - ProAgent: Building Proactive Cooperative Agents with Large Language
Models [89.53040828210945]
ProAgent is a novel framework that harnesses large language models to create proactive agents.
ProAgent can analyze the present state, and infer the intentions of teammates from observations.
ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various coordination scenarios.
arXiv Detail & Related papers (2023-08-22T10:36:56Z) - Towards Few-shot Coordination: Revisiting Ad-hoc Teamplay Challenge In
the Game of Hanabi [15.917861586043813]
We show that state-of-the-art ZSC algorithms have poor performance when paired with agents trained with different learning methods.
We create a framework based on a popular cooperative multi-agent game called Hanabi to evaluate the adaptability of MARL methods.
arXiv Detail & Related papers (2023-08-20T14:44:50Z) - Learning in Cooperative Multiagent Systems Using Cognitive and Machine
Models [1.0742675209112622]
Multi-Agent Systems (MAS) are critical for many applications requiring collaboration and coordination with humans.
One major challenge is the simultaneous learning and interaction of independent agents in dynamic environments.
We propose three variants of Multi-Agent IBL models (MAIBL)
We demonstrate that the MAIBL models exhibit faster learning and achieve better coordination in a dynamic CMOTP task with various settings of rewards compared to current MADRL models.
arXiv Detail & Related papers (2023-08-18T00:39:06Z) - Learning Reward Machines in Cooperative Multi-Agent Tasks [75.79805204646428]
This paper presents a novel approach to Multi-Agent Reinforcement Learning (MARL)
It combines cooperative task decomposition with the learning of reward machines (RMs) encoding the structure of the sub-tasks.
The proposed method helps deal with the non-Markovian nature of the rewards in partially observable environments.
arXiv Detail & Related papers (2023-03-24T15:12:28Z) - Emergence of Theory of Mind Collaboration in Multiagent Systems [65.97255691640561]
We propose an adaptive training algorithm to develop effective collaboration between agents with ToM.
We evaluate our algorithms with two games, where our algorithm surpasses all previous decentralized execution algorithms without modeling ToM.
arXiv Detail & Related papers (2021-09-30T23:28:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.