No-Regret Learning in Games is Turing Complete
- URL: http://arxiv.org/abs/2202.11871v1
- Date: Thu, 24 Feb 2022 02:37:50 GMT
- Title: No-Regret Learning in Games is Turing Complete
- Authors: Gabriel P. Andrade, Rafael Frongillo, Georgios Piliouras
- Abstract summary: We prove Turing completeness of the replicator dynamic on matrix games, one of the simplest possible settings.
Our results imply the undecicability of reachability problems for learning algorithms in games, a special case of which is determining equilibrium convergence.
- Score: 33.03065693224728
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Games are natural models for multi-agent machine learning settings, such as
generative adversarial networks (GANs). The desirable outcomes from algorithmic
interactions in these games are encoded as game theoretic equilibrium concepts,
e.g. Nash and coarse correlated equilibria. As directly computing an
equilibrium is typically impractical, one often aims to design learning
algorithms that iteratively converge to equilibria. A growing body of negative
results casts doubt on this goal, from non-convergence to chaotic and even
arbitrary behaviour. In this paper we add a strong negative result to this
list: learning in games is Turing complete. Specifically, we prove Turing
completeness of the replicator dynamic on matrix games, one of the simplest
possible settings. Our results imply the undecicability of reachability
problems for learning algorithms in games, a special case of which is
determining equilibrium convergence.
Related papers
- Neural Population Learning beyond Symmetric Zero-sum Games [52.20454809055356]
We introduce NeuPL-JPSRO, a neural population learning algorithm that benefits from transfer learning of skills and converges to a Coarse Correlated (CCE) of the game.
Our work shows that equilibrium convergent population learning can be implemented at scale and in generality.
arXiv Detail & Related papers (2024-01-10T12:56:24Z) - Hardness of Independent Learning and Sparse Equilibrium Computation in
Markov Games [70.19141208203227]
We consider the problem of decentralized multi-agent reinforcement learning in Markov games.
We show that no algorithm attains no-regret in general-sum games when executed independently by all players.
We show that our lower bounds hold even for seemingly easier setting in which all agents are controlled by a centralized algorithm.
arXiv Detail & Related papers (2023-03-22T03:28:12Z) - On the Convergence of No-Regret Learning Dynamics in Time-Varying Games [89.96815099996132]
We characterize the convergence of optimistic gradient descent (OGD) in time-varying games.
Our framework yields sharp convergence bounds for the equilibrium gap of OGD in zero-sum games.
We also provide new insights on dynamic regret guarantees in static games.
arXiv Detail & Related papers (2023-01-26T17:25:45Z) - Learning Correlated Equilibria in Mean-Field Games [62.14589406821103]
We develop the concepts of Mean-Field correlated and coarse-correlated equilibria.
We show that they can be efficiently learnt in emphall games, without requiring any additional assumption on the structure of the game.
arXiv Detail & Related papers (2022-08-22T08:31:46Z) - No-Regret Learning in Time-Varying Zero-Sum Games [99.86860277006318]
Learning from repeated play in a fixed zero-sum game is a classic problem in game theory and online learning.
We develop a single parameter-free algorithm that simultaneously enjoys favorable guarantees under three performance measures.
Our algorithm is based on a two-layer structure with a meta-algorithm learning over a group of black-box base-learners satisfying a certain property.
arXiv Detail & Related papers (2022-01-30T06:10:04Z) - Multiplayer Performative Prediction: Learning in Decision-Dependent
Games [18.386569111954213]
This paper formulates a new game theoretic framework for multi-player performative prediction.
We focus on two distinct solution concepts, namely (i) performatively stable equilibria and (ii) Nash equilibria of the game.
We show that under mild assumptions, the performatively stable equilibria can be found efficiently by a variety of algorithms.
arXiv Detail & Related papers (2022-01-10T15:31:10Z) - Towards convergence to Nash equilibria in two-team zero-sum games [17.4461045395989]
Two-team zero-sum games are defined as multi-player games where players are split into two competing sets of agents.
We focus on the solution concept of Nash equilibria (NE)
We show that computing NE for this class of games is $textithard$ for the complexity class $mathrm$.
arXiv Detail & Related papers (2021-11-07T21:15:35Z) - Learning in Matrix Games can be Arbitrarily Complex [24.6667169480096]
We show that replicator dynamics is rich enough to be able to approximate arbitrary dynamical systems.
We show how replicator dynamics can effectively reproduce the well-known strange attractor of Lonrenz dynamics.
arXiv Detail & Related papers (2021-03-05T00:41:35Z) - Hindsight and Sequential Rationality of Correlated Play [18.176128899338433]
We look at algorithms that ensure strong performance in hindsight relative to what could have been achieved with modified behavior.
We develop and advocate for this hindsight framing of learning in general sequential decision-making settings.
We present examples illustrating the distinct strengths and weaknesses of each type of equilibrium in the literature.
arXiv Detail & Related papers (2020-12-10T18:30:21Z) - No-regret learning and mixed Nash equilibria: They do not mix [64.37511607254115]
We study the dynamics of "follow-the-regularized-leader" (FTRL)
We show that any Nash equilibrium which is not strict cannot be stable and attracting under FTRL.
This result has significant implications for predicting the outcome of a learning process.
arXiv Detail & Related papers (2020-10-19T13:49:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.