Convergence of Deep Fictitious Play for Stochastic Differential Games
- URL: http://arxiv.org/abs/2008.05519v2
- Date: Sun, 21 Mar 2021 07:46:08 GMT
- Title: Convergence of Deep Fictitious Play for Stochastic Differential Games
- Authors: Jiequn Han, Ruimeng Hu, Jihao Long
- Abstract summary: Recently proposed machine learning algorithm, deep fictitious play, provides a novel efficient tool for finding Markovian Nash equilibrium of large $N$ asymmetric differential games.
By incorporating the idea of fictitious play, the algorithm decouples the game into $N$ sub-optimization problems.
We show that the strategy based on DFP forms an $eps$-Nash equilibrium.
- Score: 6.875312133832078
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stochastic differential games have been used extensively to model agents'
competitions in Finance, for instance, in P2P lending platforms from the
Fintech industry, the banking system for systemic risk, and insurance markets.
The recently proposed machine learning algorithm, deep fictitious play,
provides a novel efficient tool for finding Markovian Nash equilibrium of large
$N$-player asymmetric stochastic differential games [J. Han and R. Hu,
Mathematical and Scientific Machine Learning Conference, pages 221-245, PMLR,
2020]. By incorporating the idea of fictitious play, the algorithm decouples
the game into $N$ sub-optimization problems, and identifies each player's
optimal strategy with the deep backward stochastic differential equation (BSDE)
method parallelly and repeatedly. In this paper, we prove the convergence of
deep fictitious play (DFP) to the true Nash equilibrium. We can also show that
the strategy based on DFP forms an $\eps$-Nash equilibrium. We generalize the
algorithm by proposing a new approach to decouple the games, and present
numerical results of large population games showing the empirical convergence
of the algorithm beyond the technical assumptions in the theorems.
Related papers
- Last-Iterate Convergence of Payoff-Based Independent Learning in Zero-Sum Stochastic Games [31.554420227087043]
We develop learning dynamics that are payoff-based, convergent, rational, and symmetric between the two players.
In the matrix game setting, the results imply a complexity of $O(epsilon-1)$ to find the Nash distribution.
In the game setting, the results also imply a complexity of $O(epsilon-8)$ to find a Nash equilibrium.
arXiv Detail & Related papers (2024-09-02T20:07:25Z) - On Tractable $Φ$-Equilibria in Non-Concave Games [53.212133025684224]
Non-concave games introduce significant game-theoretic and optimization challenges.
We show that when $Phi$ is finite, there exists an efficient uncoupled learning algorithm that converges to the corresponding $Phi$-equilibria.
We also show that Online Gradient Descent can efficiently approximate $Phi$-equilibria in non-trivial regimes.
arXiv Detail & Related papers (2024-03-13T01:51:30Z) - Optimistic Policy Gradient in Multi-Player Markov Games with a Single
Controller: Convergence Beyond the Minty Property [89.96815099996132]
We develop a new framework to characterize optimistic policy gradient methods in multi-player games with a single controller.
Our approach relies on a natural generalization of the classical Minty property that we introduce, which we anticipate to have further applications beyond Markov games.
arXiv Detail & Related papers (2023-12-19T11:34:10Z) - Scalable and Independent Learning of Nash Equilibrium Policies in
$n$-Player Stochastic Games with Unknown Independent Chains [1.0878040851638]
We study games with independent chains and unknown transition matrices.
In this class of games, players control their own internal Markov chains whose transitions do not depend on the states/actions of other players.
We propose a fully decentralized mirror descent algorithm to learn an $epsilon$-NE policy.
arXiv Detail & Related papers (2023-12-04T03:04:09Z) - Representation Learning for General-sum Low-rank Markov Games [63.119870889883224]
We study multi-agent general-sum Markov games with nonlinear function approximation.
We focus on low-rank Markov games whose transition matrix admits a hidden low-rank structure on top of an unknown non-linear representation.
arXiv Detail & Related papers (2022-10-30T22:58:22Z) - Minimax-Optimal Multi-Agent RL in Zero-Sum Markov Games With a
Generative Model [50.38446482252857]
Two-player zero-sum Markov games are arguably the most basic setting in multi-agent reinforcement learning.
We develop a learning algorithm that learns an $varepsilon$-approximate Markov NE policy using $$ widetildeObigg.
We derive a refined regret bound for FTRL that makes explicit the role of variance-type quantities.
arXiv Detail & Related papers (2022-08-22T17:24:55Z) - Learning Correlated Equilibria in Mean-Field Games [62.14589406821103]
We develop the concepts of Mean-Field correlated and coarse-correlated equilibria.
We show that they can be efficiently learnt in emphall games, without requiring any additional assumption on the structure of the game.
arXiv Detail & Related papers (2022-08-22T08:31:46Z) - On the Convergence of Fictitious Play: A Decomposition Approach [17.607284715519587]
We extend the convergence results of Fictitious Play (FP) to the combinations of such games and beyond.
We develop a linear relationship unifying cooperation and competition in the sense that these two classes of games are mutually transferable.
We analyze a non-convergent example of FP, the Shapley game, and develop sufficient conditions for FP to converge.
arXiv Detail & Related papers (2022-05-03T13:04:09Z) - Towards convergence to Nash equilibria in two-team zero-sum games [17.4461045395989]
Two-team zero-sum games are defined as multi-player games where players are split into two competing sets of agents.
We focus on the solution concept of Nash equilibria (NE)
We show that computing NE for this class of games is $textithard$ for the complexity class $mathrm$.
arXiv Detail & Related papers (2021-11-07T21:15:35Z) - Learning Zero-Sum Simultaneous-Move Markov Games Using Function
Approximation and Correlated Equilibrium [116.56359444619441]
We develop provably efficient reinforcement learning algorithms for two-player zero-sum finite-horizon Markov games.
In the offline setting, we control both players and aim to find the Nash Equilibrium by minimizing the duality gap.
In the online setting, we control a single player playing against an arbitrary opponent and aim to minimize the regret.
arXiv Detail & Related papers (2020-02-17T17:04:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.