No-regret learning and mixed Nash equilibria: They do not mix
- URL: http://arxiv.org/abs/2010.09514v2
- Date: Tue, 20 Oct 2020 06:15:27 GMT
- Title: No-regret learning and mixed Nash equilibria: They do not mix
- Authors: Lampros Flokas and Emmanouil-Vasileios Vlatakis-Gkaragkounis and
Thanasis Lianeas and Panayotis Mertikopoulos and Georgios Piliouras
- Abstract summary: We study the dynamics of "follow-the-regularized-leader" (FTRL)
We show that any Nash equilibrium which is not strict cannot be stable and attracting under FTRL.
This result has significant implications for predicting the outcome of a learning process.
- Score: 64.37511607254115
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding the behavior of no-regret dynamics in general $N$-player games
is a fundamental question in online learning and game theory. A folk result in
the field states that, in finite games, the empirical frequency of play under
no-regret learning converges to the game's set of coarse correlated equilibria.
By contrast, our understanding of how the day-to-day behavior of the dynamics
correlates to the game's Nash equilibria is much more limited, and only partial
results are known for certain classes of games (such as zero-sum or congestion
games). In this paper, we study the dynamics of "follow-the-regularized-leader"
(FTRL), arguably the most well-studied class of no-regret dynamics, and we
establish a sweeping negative result showing that the notion of mixed Nash
equilibrium is antithetical to no-regret learning. Specifically, we show that
any Nash equilibrium which is not strict (in that every player has a unique
best response) cannot be stable and attracting under the dynamics of FTRL. This
result has significant implications for predicting the outcome of a learning
process as it shows unequivocally that only strict (and hence, pure) Nash
equilibria can emerge as stable limit points thereof.
Related papers
- No-Regret Learning of Nash Equilibrium for Black-Box Games via Gaussian Processes [11.846329468283583]
This paper investigates the challenge of learning in black-box games, where the underlying utility function is unknown to any of the agents.
We provide a no-regret learning algorithm that utilizes Gaussian processes to identify the equilibrium in such games.
arXiv Detail & Related papers (2024-05-14T04:58:23Z) - On Tractable $Φ$-Equilibria in Non-Concave Games [53.212133025684224]
Non-concave games introduce significant game-theoretic and optimization challenges.
We show that when $Phi$ is finite, there exists an efficient uncoupled learning algorithm that converges to the corresponding $Phi$-equilibria.
We also show that Online Gradient Descent can efficiently approximate $Phi$-equilibria in non-trivial regimes.
arXiv Detail & Related papers (2024-03-13T01:51:30Z) - A Black-box Approach for Non-stationary Multi-agent Reinforcement Learning [53.83345471268163]
We investigate learning the equilibria in non-stationary multi-agent systems.
We show how to test for various types of equilibria by a black-box reduction to single-agent learning.
arXiv Detail & Related papers (2023-06-12T23:48:24Z) - Abstracting Imperfect Information Away from Two-Player Zero-Sum Games [85.27865680662973]
Nayyar et al. (2013) showed that imperfect information can be abstracted away from common-payoff games by having players publicly announce their policies as they play.
This work shows that certain regularized equilibria do not possess the aforementioned non-correspondence problem.
Because these regularized equilibria can be made arbitrarily close to Nash equilibria, our result opens the door to a new perspective to solving two-player zero-sum games.
arXiv Detail & Related papers (2023-01-22T16:54:06Z) - Nash Equilibria and Pitfalls of Adversarial Training in Adversarial
Robustness Games [51.90475640044073]
We study adversarial training as an alternating best-response strategy in a 2-player zero-sum game.
On the other hand, a unique pure Nash equilibrium of the game exists and is provably robust.
arXiv Detail & Related papers (2022-10-23T03:21:01Z) - Nash, Conley, and Computation: Impossibility and Incompleteness in Game
Dynamics [28.815822236291392]
We show that no game dynamics can converge to $epsilon$-Nash equilibria.
We also prove a stronger result for $epsilon$-approximate Nash equilibria.
arXiv Detail & Related papers (2022-03-26T18:27:40Z) - No-Regret Learning in Games is Turing Complete [33.03065693224728]
We prove Turing completeness of the replicator dynamic on matrix games, one of the simplest possible settings.
Our results imply the undecicability of reachability problems for learning algorithms in games, a special case of which is determining equilibrium convergence.
arXiv Detail & Related papers (2022-02-24T02:37:50Z) - Multiplayer Performative Prediction: Learning in Decision-Dependent
Games [18.386569111954213]
This paper formulates a new game theoretic framework for multi-player performative prediction.
We focus on two distinct solution concepts, namely (i) performatively stable equilibria and (ii) Nash equilibria of the game.
We show that under mild assumptions, the performatively stable equilibria can be found efficiently by a variety of algorithms.
arXiv Detail & Related papers (2022-01-10T15:31:10Z) - Sample-Efficient Learning of Stackelberg Equilibria in General-Sum Games [78.65798135008419]
It remains vastly open how to learn the Stackelberg equilibrium in general-sum games efficiently from samples.
This paper initiates the theoretical study of sample-efficient learning of the Stackelberg equilibrium in two-player turn-based general-sum games.
arXiv Detail & Related papers (2021-02-23T05:11:07Z) - Survival of the strictest: Stable and unstable equilibria under
regularized learning with partial information [32.384868685390906]
We examine the Nash equilibrium convergence properties of no-regret learning in general N-player games.
We establish a comprehensive equivalence between the stability of a Nash equilibrium and its support.
It provides a clear refinement criterion for the prediction of the day-to-day behavior of no-regret learning in games.
arXiv Detail & Related papers (2021-01-12T18:55:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.