Approximately Solving Mean Field Games via Entropy-Regularized Deep
Reinforcement Learning
- URL: http://arxiv.org/abs/2102.01585v1
- Date: Tue, 2 Feb 2021 16:22:07 GMT
- Title: Approximately Solving Mean Field Games via Entropy-Regularized Deep
Reinforcement Learning
- Authors: Kai Cui, Heinz Koeppl
- Abstract summary: We show that all discrete-time finite MFGs with non-constant fixed point operators fail to be contractive as typically assumed in existing MFG literature.
We obtain provable convergence to approximate fixed points where existing methods fail, and reach the original goal of approximate Nash equilibria.
- Score: 33.77849245250632
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recent mean field game (MFG) formalism facilitates otherwise intractable
computation of approximate Nash equilibria in many-agent settings. In this
paper, we consider discrete-time finite MFGs subject to finite-horizon
objectives. We show that all discrete-time finite MFGs with non-constant fixed
point operators fail to be contractive as typically assumed in existing MFG
literature, barring convergence via fixed point iteration. Instead, we
incorporate entropy-regularization and Boltzmann policies into the fixed point
iteration. As a result, we obtain provable convergence to approximate fixed
points where existing methods fail, and reach the original goal of approximate
Nash equilibria. All proposed methods are evaluated with respect to their
exploitability, on both instructive examples with tractable exact solutions and
high-dimensional problems where exact methods become intractable. In
high-dimensional scenarios, we apply established deep reinforcement learning
methods and empirically combine fictitious play with our approximations.
Related papers
- Last Iterate Convergence in Monotone Mean Field Games [5.407319151576265]
Mean Field Game (MFG) is a framework utilized to model and approximate the behavior of a large number of agents.
We propose the use of a simple, proximal-point-type algorithm to compute equilibria for MFGs.
We provide the first last-iterate convergence guarantee under the Lasry--Lions-type monotonicity condition.
arXiv Detail & Related papers (2024-10-07T15:28:18Z) - A Unified Theory of Stochastic Proximal Point Methods without Smoothness [52.30944052987393]
Proximal point methods have attracted considerable interest owing to their numerical stability and robustness against imperfect tuning.
This paper presents a comprehensive analysis of a broad range of variations of the proximal point method (SPPM)
arXiv Detail & Related papers (2024-05-24T21:09:19Z) - Maximum Causal Entropy Inverse Reinforcement Learning for Mean-Field
Games [3.2228025627337864]
We introduce the casual entropy Inverse Reinforcement (IRL) problem for discrete-time mean-field games (MFGs) under an infinite-horizon discounted-reward optimality criterion.
We present by formulating the MFG problem as a generalized Nash equilibrium problem (GN), which is capable of computing the meanfield equilibrium problem.
This method is employed to produce data for a numerical example.
arXiv Detail & Related papers (2024-01-12T13:22:03Z) - An Exponentially Converging Particle Method for the Mixed Nash
Equilibrium of Continuous Games [0.0]
We consider the problem of computing mixed Nash equilibria of two-player zero-sum games with continuous sets of pure strategies and with first-order access to the payoff function.
This problem arises for example in game-inspired machine learning applications, such as distributionally-robust learning.
We introduce and analyze a particle-based method that enjoys guaranteed local convergence for this problem.
arXiv Detail & Related papers (2022-11-02T17:03:40Z) - A unified stochastic approximation framework for learning in games [82.74514886461257]
We develop a flexible approximation framework for analyzing the long-run behavior of learning in games (both continuous and finite)
The proposed analysis template incorporates a wide array of popular learning algorithms, including gradient-based methods, exponential/multiplicative weights for learning in finite games, optimistic and bandit variants of the above, etc.
arXiv Detail & Related papers (2022-06-08T14:30:38Z) - Beyond Exact Gradients: Convergence of Stochastic Soft-Max Policy Gradient Methods with Entropy Regularization [20.651913793555163]
We revisit the classical entropy regularized policy gradient methods with the soft-max policy parametrization.
We establish a global optimality convergence result and a sample complexity of $widetildemathcalO(frac1epsilon2)$ for the proposed algorithm.
arXiv Detail & Related papers (2021-10-19T17:21:09Z) - Stochastic Gradient Descent-Ascent and Consensus Optimization for Smooth
Games: Convergence Analysis under Expected Co-coercivity [49.66890309455787]
We introduce the expected co-coercivity condition, explain its benefits, and provide the first last-iterate convergence guarantees of SGDA and SCO.
We prove linear convergence of both methods to a neighborhood of the solution when they use constant step-size.
Our convergence guarantees hold under the arbitrary sampling paradigm, and we give insights into the complexity of minibatching.
arXiv Detail & Related papers (2021-06-30T18:32:46Z) - On the Convergence of Stochastic Extragradient for Bilinear Games with
Restarted Iteration Averaging [96.13485146617322]
We present an analysis of the ExtraGradient (SEG) method with constant step size, and present variations of the method that yield favorable convergence.
We prove that when augmented with averaging, SEG provably converges to the Nash equilibrium, and such a rate is provably accelerated by incorporating a scheduled restarting procedure.
arXiv Detail & Related papers (2021-06-30T17:51:36Z) - Global Convergence of Policy Gradient for Linear-Quadratic Mean-Field
Control/Game in Continuous Time [109.06623773924737]
We study the policy gradient method for the linear-quadratic mean-field control and game.
We show that it converges to the optimal solution at a linear rate, which is verified by a synthetic simulation.
arXiv Detail & Related papers (2020-08-16T06:34:11Z) - Tackling the Objective Inconsistency Problem in Heterogeneous Federated
Optimization [93.78811018928583]
This paper provides a framework to analyze the convergence of federated heterogeneous optimization algorithms.
We propose FedNova, a normalized averaging method that eliminates objective inconsistency while preserving fast error convergence.
arXiv Detail & Related papers (2020-07-15T05:01:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.