Global Nash Equilibrium in Non-convex Multi-player Game: Theory and
Algorithms
- URL: http://arxiv.org/abs/2301.08015v1
- Date: Thu, 19 Jan 2023 11:36:50 GMT
- Title: Global Nash Equilibrium in Non-convex Multi-player Game: Theory and
Algorithms
- Authors: Guanpu Chen, Gehui Xu, Fengxiang He, Yiguang Hong, Leszek Rutkowski,
and Dacheng Tao
- Abstract summary: We show that Nash equilibrium (NE) is acceptable to all players in a multi-player game.
We also show that no one can benefit unilaterally from the general theory step by step.
- Score: 66.8634598612777
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Wide machine learning tasks can be formulated as non-convex multi-player
games, where Nash equilibrium (NE) is an acceptable solution to all players,
since no one can benefit from changing its strategy unilaterally. Attributed to
the non-convexity, obtaining the existence condition of global NE is
challenging, let alone designing theoretically guaranteed realization
algorithms. This paper takes conjugate transformation to the formulation of
non-convex multi-player games, and casts the complementary problem into a
variational inequality (VI) problem with a continuous pseudo-gradient mapping.
We then prove the existence condition of global NE: the solution to the VI
problem satisfies a duality relation. Based on this VI formulation, we design a
conjugate-based ordinary differential equation (ODE) to approach global NE,
which is proved to have an exponential convergence rate. To make the dynamics
more implementable, we further derive a discretized algorithm. We apply our
algorithm to two typical scenarios: multi-player generalized monotone game and
multi-player potential game. In the two settings, we prove that the step-size
setting is required to be $\mathcal{O}(1/k)$ and $\mathcal{O}(1/\sqrt k)$ to
yield the convergence rates of $\mathcal{O}(1/ k)$ and $\mathcal{O}(1/\sqrt
k)$, respectively. Extensive experiments in robust neural network training and
sensor localization are in full agreement with our theory.
Related papers
- Fast Last-Iterate Convergence of Learning in Games Requires Forgetful Algorithms [71.73971094342349]
Self-play via online learning is one of the premier ways to solve large-scale two-player zero-sum games.
We show that OMWU offers several advantages including logarithmic dependence on the size of the payoff matrix.
We prove that a broad class of algorithms that do not forget the past quickly all suffer the same issue.
arXiv Detail & Related papers (2024-06-15T13:26:17Z) - MF-OML: Online Mean-Field Reinforcement Learning with Occupation Measures for Large Population Games [5.778024594615575]
This paper proposes an online mean-field reinforcement learning algorithm for computing Nash equilibria of sequential games.
MFOML is the first fully approximate multi-agent reinforcement learning algorithm for provably solving Nash equilibria.
As a byproduct, we also obtain the first tractable globally convergent computational for approximate computing of monotone mean-field games.
arXiv Detail & Related papers (2024-05-01T02:19:31Z) - Distributed Extra-gradient with Optimal Complexity and Communication
Guarantees [60.571030754252824]
We consider monotone variational inequality (VI) problems in multi-GPU settings where multiple processors/workers/clients have access to local dual vectors.
Extra-gradient, which is a de facto algorithm for monotone VI problems, has not been designed to be communication-efficient.
We propose a quantized generalized extra-gradient (Q-GenX), which is an unbiased and adaptive compression method tailored to solve VIs.
arXiv Detail & Related papers (2023-08-17T21:15:04Z) - Learning in Multi-Player Stochastic Games [1.0878040851638]
We consider the problem of simultaneous learning in games with many players in the finite-horizon setting.
While the typical target solution for a game is a Nash equilibrium, this is intractable with many players.
We turn to a different target: algorithms which generate an equilibrium when they are used by all players.
arXiv Detail & Related papers (2022-10-25T19:02:03Z) - Minimax-Optimal Multi-Agent RL in Zero-Sum Markov Games With a
Generative Model [50.38446482252857]
Two-player zero-sum Markov games are arguably the most basic setting in multi-agent reinforcement learning.
We develop a learning algorithm that learns an $varepsilon$-approximate Markov NE policy using $$ widetildeObigg.
We derive a refined regret bound for FTRL that makes explicit the role of variance-type quantities.
arXiv Detail & Related papers (2022-08-22T17:24:55Z) - Learning Two-Player Mixture Markov Games: Kernel Function Approximation
and Correlated Equilibrium [157.0902680672422]
We consider learning Nash equilibria in two-player zero-sum Markov Games with nonlinear function approximation.
We propose a novel online learning algorithm to find a Nash equilibrium by minimizing the duality gap.
arXiv Detail & Related papers (2022-08-10T14:21:54Z) - Towards convergence to Nash equilibria in two-team zero-sum games [17.4461045395989]
Two-team zero-sum games are defined as multi-player games where players are split into two competing sets of agents.
We focus on the solution concept of Nash equilibria (NE)
We show that computing NE for this class of games is $textithard$ for the complexity class $mathrm$.
arXiv Detail & Related papers (2021-11-07T21:15:35Z) - On the Impossibility of Global Convergence in Multi-Loss Optimization [10.081768261298098]
We prove that desirable convergence properties cannot simultaneously hold for any algorithm.
Our result has more to do with the existence of games with no satisfactory outcomes, than with algorithms per se.
It remains an open question whether such behavior can arise in high-dimensional games of interest to ML practitioners.
arXiv Detail & Related papers (2020-05-26T12:11:18Z) - Learning Zero-Sum Simultaneous-Move Markov Games Using Function
Approximation and Correlated Equilibrium [116.56359444619441]
We develop provably efficient reinforcement learning algorithms for two-player zero-sum finite-horizon Markov games.
In the offline setting, we control both players and aim to find the Nash Equilibrium by minimizing the duality gap.
In the online setting, we control a single player playing against an arbitrary opponent and aim to minimize the regret.
arXiv Detail & Related papers (2020-02-17T17:04:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.