Game-Theoretical Perspectives on Active Equilibria: A Preferred Solution
Concept over Nash Equilibria
- URL: http://arxiv.org/abs/2210.16175v1
- Date: Fri, 28 Oct 2022 14:45:39 GMT
- Title: Game-Theoretical Perspectives on Active Equilibria: A Preferred Solution
Concept over Nash Equilibria
- Authors: Dong-Ki Kim, Matthew Riemer, Miao Liu, Jakob N. Foerster, Gerald
Tesauro, Jonathan P. How
- Abstract summary: An effective approach in multiagent reinforcement learning is to consider the learning process of agents and influence their future policies.
This new solution concept is general such that standard solution concepts, such as a Nash equilibrium, are special cases of active equilibria.
We analyze active equilibria from a game-theoretic perspective by closely studying examples where Nash equilibria are known.
- Score: 61.093297204685264
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multiagent learning settings are inherently more difficult than single-agent
learning because each agent interacts with other simultaneously learning agents
in a shared environment. An effective approach in multiagent reinforcement
learning is to consider the learning process of agents and influence their
future policies toward desirable behaviors from each agent's perspective.
Importantly, if each agent maximizes its long-term rewards by accounting for
the impact of its behavior on the set of convergence policies, the resulting
multiagent system reaches an active equilibrium. While this new solution
concept is general such that standard solution concepts, such as a Nash
equilibrium, are special cases of active equilibria, it is unclear when an
active equilibrium is a preferred equilibrium over other solution concepts. In
this paper, we analyze active equilibria from a game-theoretic perspective by
closely studying examples where Nash equilibria are known. By directly
comparing active equilibria to Nash equilibria in these examples, we find that
active equilibria find more effective solutions than Nash equilibria,
concluding that an active equilibrium is the desired solution for multiagent
learning settings.
Related papers
- A Black-box Approach for Non-stationary Multi-agent Reinforcement Learning [53.83345471268163]
We investigate learning the equilibria in non-stationary multi-agent systems.
We show how to test for various types of equilibria by a black-box reduction to single-agent learning.
arXiv Detail & Related papers (2023-06-12T23:48:24Z) - Inducing Stackelberg Equilibrium through Spatio-Temporal Sequential
Decision-Making in Multi-Agent Reinforcement Learning [17.101534531286298]
We construct a Nash-level policy model based on a conditional hypernetwork shared by all agents.
This approach allows for asymmetric training with symmetric execution, with each agent responding optimally conditioned on the decisions made by superior agents.
Experiments demonstrate that our method effectively converges to the SE policies in repeated matrix game scenarios.
arXiv Detail & Related papers (2023-04-20T14:47:54Z) - Uncoupled Learning of Differential Stackelberg Equilibria with Commitments [43.098826226730246]
We present uncoupled'' learning dynamics based on zeroth-order gradient estimators.
We prove that they converge to differential Stackelberg equilibria under the same conditions as previous coupled methods.
We also present an online mechanism by which symmetric learners can negotiate leader-follower roles.
arXiv Detail & Related papers (2023-02-07T12:46:54Z) - Pareto Actor-Critic for Equilibrium Selection in Multi-Agent
Reinforcement Learning [18.20664209675016]
This work focuses on equilibrium selection in no-conflict multi-agent games.
Pareto Actor-Critic (Pareto-AC) is an actor-critic algorithm that maximises the returns of all agents.
arXiv Detail & Related papers (2022-09-28T18:14:34Z) - Efficient Model-based Multi-agent Reinforcement Learning via Optimistic
Equilibrium Computation [93.52573037053449]
H-MARL (Hallucinated Multi-Agent Reinforcement Learning) learns successful equilibrium policies after a few interactions with the environment.
We demonstrate our approach experimentally on an autonomous driving simulation benchmark.
arXiv Detail & Related papers (2022-03-14T17:24:03Z) - Finding General Equilibria in Many-Agent Economic Simulations Using Deep
Reinforcement Learning [72.23843557783533]
We show that deep reinforcement learning can discover stable solutions that are epsilon-Nash equilibria for a meta-game over agent types.
Our approach is more flexible and does not need unrealistic assumptions, e.g., market clearing.
We demonstrate our approach in real-business-cycle models, a representative family of DGE models, with 100 worker-consumers, 10 firms, and a government who taxes and redistributes.
arXiv Detail & Related papers (2022-01-03T17:00:17Z) - Sample-Efficient Learning of Stackelberg Equilibria in General-Sum Games [78.65798135008419]
It remains vastly open how to learn the Stackelberg equilibrium in general-sum games efficiently from samples.
This paper initiates the theoretical study of sample-efficient learning of the Stackelberg equilibrium in two-player turn-based general-sum games.
arXiv Detail & Related papers (2021-02-23T05:11:07Z) - On Information Asymmetry in Competitive Multi-Agent Reinforcement
Learning: Convergence and Optimality [78.76529463321374]
We study the system of interacting non-cooperative two Q-learning agents.
We show that this information asymmetry can lead to a stable outcome of population learning.
arXiv Detail & Related papers (2020-10-21T11:19:53Z) - Calibration of Shared Equilibria in General Sum Partially Observable
Markov Games [15.572157454411533]
We consider a general sum partially observable Markov game where agents of different types share a single policy network.
This paper aims at i) formally understanding equilibria reached by such agents, and ii) matching emergent phenomena of such equilibria to real-world targets.
arXiv Detail & Related papers (2020-06-23T15:14:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.