Kernelized Multiplicative Weights for 0/1-Polyhedral Games: Bridging the
Gap Between Learning in Extensive-Form and Normal-Form Games
- URL: http://arxiv.org/abs/2202.00237v1
- Date: Tue, 1 Feb 2022 06:28:51 GMT
- Title: Kernelized Multiplicative Weights for 0/1-Polyhedral Games: Bridging the
Gap Between Learning in Extensive-Form and Normal-Form Games
- Authors: Gabriele Farina, Chung-Wei Lee, Haipeng Luo, Christian Kroer
- Abstract summary: We show that the Optimistic Multiplicative Weights Update (OMWU) algorithm can be simulated on the normal-form equivalent of an EFG in linear time per iteration in the game tree size using a kernel trick.
Specifically, KOMWU gives the first algorithm that guarantees at the same time last-iterate convergence.
- Score: 76.21916750766277
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While extensive-form games (EFGs) can be converted into normal-form games
(NFGs), doing so comes at the cost of an exponential blowup of the strategy
space. So, progress on NFGs and EFGs has historically followed separate tracks,
with the EFG community often having to catch up with advances (e.g.,
last-iterate convergence and predictive regret bounds) from the larger NFG
community. In this paper we show that the Optimistic Multiplicative Weights
Update (OMWU) algorithm -- the premier learning algorithm for NFGs -- can be
simulated on the normal-form equivalent of an EFG in linear time per iteration
in the game tree size using a kernel trick. The resulting algorithm, Kernelized
OMWU (KOMWU), applies more broadly to all convex games whose strategy space is
a polytope with 0/1 integral vertices, as long as the kernel can be evaluated
efficiently. In the particular case of EFGs, KOMWU closes several standing gaps
between NFG and EFG learning, by enabling direct, black-box transfer to EFGs of
desirable properties of learning dynamics that were so far known to be
achievable only in NFGs. Specifically, KOMWU gives the first algorithm that
guarantees at the same time last-iterate convergence, lower dependence on the
size of the game tree than all prior algorithms, and $\tilde{\mathcal{O}}(1)$
regret when followed by all players.
Related papers
- Learning Discrete-Time Major-Minor Mean Field Games [61.09249862334384]
We propose a novel discrete time version of major-minor MFGs (M3FGs) and a learning algorithm based on fictitious play and partitioning the probability simplex.
M3FGs generalize MFGs with common noise and can handle not only random exogeneous environment states but also major players.
arXiv Detail & Related papers (2023-12-17T18:22:08Z) - Learning Regularized Graphon Mean-Field Games with Unknown Graphons [155.38727464526923]
We design reinforcement learning algorithms for Graphon Mean-Field Games (GMFGs)
We aim to learn the Nash Equilibrium (NE) of the regularized GMFGs when the graphons are unknown.
These algorithms are the first specifically designed for learning graphons from sampled agents.
arXiv Detail & Related papers (2023-10-26T16:19:24Z) - Learning Regularized Monotone Graphon Mean-Field Games [155.38727464526923]
We study fundamental problems in regularized Graphon Mean-Field Games (GMFGs)
We establish the existence of a Nash Equilibrium (NE) of any $lambda$-regularized GMFG.
We propose provably efficient algorithms to learn the NE in weakly monotone GMFGs.
arXiv Detail & Related papers (2023-10-12T07:34:13Z) - The Power of Regularization in Solving Extensive-Form Games [22.557806157585834]
We propose a series of new algorithms based on regularizing the payoff functions of the game.
In particular, we first show that dilated optimistic mirror descent (DOMD) can achieve a fast $tilde O(T)$ last-iterate convergence.
We also show that Reg-CFR, with a variant of optimistic mirror descent algorithm as regret-minimizer, can achieve $O(T1/4)$ best-iterate, and $O(T3/4)$ average-iterate convergence rate.
arXiv Detail & Related papers (2022-06-19T22:10:38Z) - Efficient $\Phi$-Regret Minimization in Extensive-Form Games via Online
Mirror Descent [49.93548413166884]
$Phi$-Hedge is a generic algorithm capable of learning a large class of equilibria for Normal-Form Games (NFGs)
We show that $Phi$-Hedge can be directly used to learn Nash Equilibria (zero-sum settings), Normal-Form Coarse Correlated Equilibria (NFCCE), and Extensive-Form Correlated Equilibria (EFCE) in EFGs.
We prove that, in those settings, the emph$Phi$-Hedge algorithms are equivalent to standard Mirror Descent (OMD) algorithms for
arXiv Detail & Related papers (2022-05-30T17:58:06Z) - Last-iterate Convergence in Extensive-Form Games [49.31256241275577]
We study last-iterate convergence of optimistic algorithms in sequential games.
We show that all of these algorithms enjoy last-iterate convergence, with some of them even converging exponentially fast.
arXiv Detail & Related papers (2021-06-27T22:02:26Z) - Signatured Deep Fictitious Play for Mean Field Games with Common Noise [0.0]
Existing deep learning methods for solving mean-field games (MFGs) with common noise fix the sampling common noise paths and then solve the corresponding MFGs.
We propose a novel single-loop algorithm, named signatured deep fictitious play, by which we can work with the unfixed common noise setup to avoid the nested-loop structure.
The proposed algorithm can accurately capture the effect of common uncertainty changes on mean-field equilibria without further training of neural networks.
arXiv Detail & Related papers (2021-06-06T23:09:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.