Learning to Learn Group Alignment: A Self-Tuning Credo Framework with
Multiagent Teams
- URL: http://arxiv.org/abs/2304.07337v1
- Date: Fri, 14 Apr 2023 18:16:19 GMT
- Title: Learning to Learn Group Alignment: A Self-Tuning Credo Framework with
Multiagent Teams
- Authors: David Radke and Kyle Tilbury
- Abstract summary: Mixed incentives among a population with multiagent teams has been shown to have advantages over a fully cooperative system.
We propose a framework where individual learning agents self-regulate their configuration of incentives through various parts of their reward function.
- Score: 1.370633147306388
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Mixed incentives among a population with multiagent teams has been shown to
have advantages over a fully cooperative system; however, discovering the best
mixture of incentives or team structure is a difficult and dynamic problem. We
propose a framework where individual learning agents self-regulate their
configuration of incentives through various parts of their reward function.
This work extends previous work by giving agents the ability to dynamically
update their group alignment during learning and by allowing teammates to have
different group alignment. Our model builds on ideas from hierarchical
reinforcement learning and meta-learning to learn the configuration of a reward
function that supports the development of a behavioral policy. We provide
preliminary results in a commonly studied multiagent environment and find that
agents can achieve better global outcomes by self-tuning their respective group
alignment parameters.
Related papers
- Peer Learning: Learning Complex Policies in Groups from Scratch via Action Recommendations [16.073203911932872]
Peer learning is a novel high-level reinforcement learning framework for agents learning in groups.
We show that peer learning is able to outperform single agent learning and the baseline in several challenging OpenAI Gym domains.
arXiv Detail & Related papers (2023-12-15T17:01:35Z) - ProAgent: Building Proactive Cooperative Agents with Large Language
Models [89.53040828210945]
ProAgent is a novel framework that harnesses large language models to create proactive agents.
ProAgent can analyze the present state, and infer the intentions of teammates from observations.
ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various coordination scenarios.
arXiv Detail & Related papers (2023-08-22T10:36:56Z) - AgentVerse: Facilitating Multi-Agent Collaboration and Exploring
Emergent Behaviors [93.38830440346783]
We propose a multi-agent framework framework that can collaboratively adjust its composition as a greater-than-the-sum-of-its-parts system.
Our experiments demonstrate that framework framework can effectively deploy multi-agent groups that outperform a single agent.
In view of these behaviors, we discuss some possible strategies to leverage positive ones and mitigate negative ones for improving the collaborative potential of multi-agent groups.
arXiv Detail & Related papers (2023-08-21T16:47:11Z) - ELIGN: Expectation Alignment as a Multi-Agent Intrinsic Reward [29.737986509769808]
We propose a self-supervised intrinsic reward ELIGN - expectation alignment.
Similar to how animals collaborate in a decentralized manner with those in their vicinity, agents trained with expectation alignment learn behaviors that match their neighbors' expectations.
We show that agent coordination improves through expectation alignment because agents learn to divide tasks amongst themselves, break coordination symmetries, and confuse adversaries.
arXiv Detail & Related papers (2022-10-09T22:24:44Z) - The Importance of Credo in Multiagent Learning [5.334505575267924]
We propose a model for multi-objective optimization, a credo, for agents in a system that are configured into multiple groups.
Our results indicate that the interests of teammates, or the entire system, are not required to be fully aligned for achieving globally beneficial outcomes.
arXiv Detail & Related papers (2022-04-15T14:12:13Z) - Generalization in Cooperative Multi-Agent Systems [49.16349318581611]
We study the theoretical underpinnings of Combinatorial Generalization (CG) for cooperative multi-agent systems.
CG is a highly desirable trait for autonomous systems as it can increase their utility and deployability across a wide range of applications.
arXiv Detail & Related papers (2022-01-31T21:39:56Z) - Structured Diversification Emergence via Reinforced Organization Control
and Hierarchical Consensus Learning [48.525944995851965]
We propose a structured diversification emergence MARL framework named scRochico based on reinforced organization control and hierarchical consensus learning.
scRochico is significantly better than the current SOTA algorithms in terms of exploration efficiency and cooperation strength.
arXiv Detail & Related papers (2021-02-09T11:46:12Z) - Natural Emergence of Heterogeneous Strategies in Artificially
Intelligent Competitive Teams [0.0]
We develop a competitive multi agent environment called FortAttack in which two teams compete against each other.
We observe a natural emergence of heterogeneous behavior amongst homogeneous agents when such behavior can lead to the team's success.
We propose ensemble training, in which we utilize the evolved opponent strategies to train a single policy for friendly agents.
arXiv Detail & Related papers (2020-07-06T22:35:56Z) - Learning to Incentivize Other Learning Agents [73.03133692589532]
We show how to equip RL agents with the ability to give rewards directly to other agents, using a learned incentive function.
Such agents significantly outperform standard RL and opponent-shaping agents in challenging general-sum Markov games.
Our work points toward more opportunities and challenges along the path to ensure the common good in a multi-agent future.
arXiv Detail & Related papers (2020-06-10T20:12:38Z) - Randomized Entity-wise Factorization for Multi-Agent Reinforcement
Learning [59.62721526353915]
Multi-agent settings in the real world often involve tasks with varying types and quantities of agents and non-agent entities.
Our method aims to leverage these commonalities by asking the question: What is the expected utility of each agent when only considering a randomly selected sub-group of its observed entities?''
arXiv Detail & Related papers (2020-06-07T18:28:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.