Preference Communication in Multi-Objective Normal-Form Games
- URL: http://arxiv.org/abs/2111.09191v1
- Date: Wed, 17 Nov 2021 15:30:41 GMT
- Title: Preference Communication in Multi-Objective Normal-Form Games
- Authors: Willem R\"opke, Diederik M. Roijers, Ann Now\'e, Roxana R\u{a}dulescu
- Abstract summary: We study the problem of multiple agents learning concurrently in a multi-objective environment.
We introduce four novel preference communication protocols for both cooperative and self-interested communication.
We find that preference communication can drastically alter the learning process and lead to the emergence of cyclic Nash equilibria.
- Score: 3.8099752264464883
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of multiple agents learning concurrently in a
multi-objective environment. Specifically, we consider two agents that
repeatedly play a multi-objective normal-form game. In such games, the payoffs
resulting from joint actions are vector valued. Taking a utility-based
approach, we assume a utility function exists that maps vectors to scalar
utilities and consider agents that aim to maximise the utility of expected
payoff vectors. As agents do not necessarily know their opponent's utility
function or strategy, they must learn optimal policies to interact with each
other. To aid agents in arriving at adequate solutions, we introduce four novel
preference communication protocols for both cooperative as well as
self-interested communication. Each approach describes a specific protocol for
one agent communicating preferences over their actions and how another agent
responds. These protocols are subsequently evaluated on a set of five benchmark
games against baseline agents that do not communicate. We find that preference
communication can drastically alter the learning process and lead to the
emergence of cyclic Nash equilibria which had not been previously observed in
this setting. Additionally, we introduce a communication scheme where agents
must learn when to communicate. For agents in games with Nash equilibria, we
find that communication can be beneficial but difficult to learn when agents
have different preferred equilibria. When this is not the case, agents become
indifferent to communication. In games without Nash equilibria, our results
show differences across learning rates. When using faster learners, we observe
that explicit communication becomes more prevalent at around 50% of the time,
as it helps them in learning a compromise joint policy. Slower learners retain
this pattern to a lesser degree, but show increased indifference.
Related papers
- Impact of Decentralized Learning on Player Utilities in Stackelberg Games [57.08270857260131]
In many two-agent systems, each agent learns separately and the rewards of the two agents are not perfectly aligned.
We model these systems as Stackelberg games with decentralized learning and show that standard regret benchmarks result in worst-case linear regret for at least one player.
We develop algorithms to achieve near-optimal $O(T2/3)$ regret for both players with respect to these benchmarks.
arXiv Detail & Related papers (2024-02-29T23:38:28Z) - Fast Peer Adaptation with Context-aware Exploration [63.08444527039578]
We propose a peer identification reward for learning agents in multi-agent games.
This reward motivates the agent to learn a context-aware policy for effective exploration and fast adaptation.
We evaluate our method on diverse testbeds that involve competitive (Kuhn Poker), cooperative (PO-Overcooked), or mixed (Predator-Prey-W) games with peer agents.
arXiv Detail & Related papers (2024-02-04T13:02:27Z) - ProAgent: Building Proactive Cooperative Agents with Large Language
Models [89.53040828210945]
ProAgent is a novel framework that harnesses large language models to create proactive agents.
ProAgent can analyze the present state, and infer the intentions of teammates from observations.
ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various coordination scenarios.
arXiv Detail & Related papers (2023-08-22T10:36:56Z) - Over-communicate no more: Situated RL agents learn concise communication
protocols [78.28898217947467]
It is unclear how to design artificial agents that can learn to effectively and efficiently communicate with each other.
Much research on communication emergence uses reinforcement learning (RL)
We explore situated communication in a multi-step task, where the acting agent has to forgo an environmental action to communicate.
We find that while all tested pressures can disincentivise over-communication, situated communication does it most effectively and, unlike the cost on effort, does not negatively impact emergence.
arXiv Detail & Related papers (2022-11-02T21:08:14Z) - Interpretation of Emergent Communication in Heterogeneous Collaborative
Embodied Agents [83.52684405389445]
We introduce the collaborative multi-object navigation task CoMON.
In this task, an oracle agent has detailed environment information in the form of a map.
It communicates with a navigator agent that perceives the environment visually and is tasked to find a sequence of goals.
We show that the emergent communication can be grounded to the agent observations and the spatial structure of the 3D environment.
arXiv Detail & Related papers (2021-10-12T06:56:11Z) - Implicit Communication as Minimum Entropy Coupling [42.13333133772116]
In many common-payoff games, achieving good performance requires players to develop protocols for communicating their private information implicitly.
We identify a class of partially observable common-payoff games, which we call implicit referential games, whose difficulty can be attributed to implicit communication.
We show that this method can discover performant implicit communication protocols in settings with very large spaces of messages.
arXiv Detail & Related papers (2021-07-17T17:44:30Z) - Correcting Experience Replay for Multi-Agent Communication [18.12281605882891]
We consider the problem of learning to communicate using multi-agent reinforcement learning (MARL)
A common approach is to learn off-policy, using data sampled from a replay buffer.
We introduce a 'communication correction' which accounts for the non-stationarity of observed communication induced by MARL.
arXiv Detail & Related papers (2020-10-02T20:49:24Z) - Pow-Wow: A Dataset and Study on Collaborative Communication in Pommerman [12.498028338281625]
In multi-agent learning, agents must coordinate with each other in order to succeed. For humans, this coordination is typically accomplished through the use of language.
We construct Pow-Wow, a new dataset for studying situated goal-directed human communication.
We analyze the types of communications which result in effective game strategies, annotate them accordingly, and present corpus-level statistical analysis of how trends in communications affect game outcomes.
arXiv Detail & Related papers (2020-09-13T07:11:37Z) - The Emergence of Adversarial Communication in Multi-Agent Reinforcement
Learning [6.18778092044887]
Many real-world problems require the coordination of multiple autonomous agents.
Recent work has shown the promise of Graph Neural Networks (GNNs) to learn explicit communication strategies that enable complex multi-agent coordination.
We show how a single self-interested agent is capable of learning highly manipulative communication strategies that allows it to significantly outperform a cooperative team of agents.
arXiv Detail & Related papers (2020-08-06T12:48:08Z) - Learning Individually Inferred Communication for Multi-Agent Cooperation [37.56115000150748]
We propose Individually Inferred Communication (I2C) to enable agents to learn a prior for agent-agent communication.
The prior knowledge is learned via causal inference and realized by a feed-forward neural network.
I2C can not only reduce communication overhead but also improve the performance in a variety of multi-agent cooperative scenarios.
arXiv Detail & Related papers (2020-06-11T14:07:57Z) - On Emergent Communication in Competitive Multi-Agent Teams [116.95067289206919]
We investigate whether competition for performance from an external, similar agent team could act as a social influence.
Our results show that an external competitive influence leads to improved accuracy and generalization, as well as faster emergence of communicative languages.
arXiv Detail & Related papers (2020-03-04T01:14:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.