An Exponentially Converging Particle Method for the Mixed Nash
Equilibrium of Continuous Games
- URL: http://arxiv.org/abs/2211.01280v3
- Date: Thu, 13 Apr 2023 11:56:44 GMT
- Title: An Exponentially Converging Particle Method for the Mixed Nash
Equilibrium of Continuous Games
- Authors: Guillaume Wang and L\'ena\"ic Chizat
- Abstract summary: We consider the problem of computing mixed Nash equilibria of two-player zero-sum games with continuous sets of pure strategies and with first-order access to the payoff function.
This problem arises for example in game-inspired machine learning applications, such as distributionally-robust learning.
We introduce and analyze a particle-based method that enjoys guaranteed local convergence for this problem.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of computing mixed Nash equilibria of two-player
zero-sum games with continuous sets of pure strategies and with first-order
access to the payoff function. This problem arises for example in
game-theory-inspired machine learning applications, such as
distributionally-robust learning. In those applications, the strategy sets are
high-dimensional and thus methods based on discretisation cannot tractably
return high-accuracy solutions.
In this paper, we introduce and analyze a particle-based method that enjoys
guaranteed local convergence for this problem. This method consists in
parametrizing the mixed strategies as atomic measures and applying proximal
point updates to both the atoms' weights and positions. It can be interpreted
as a time-implicit discretization of the "interacting" Wasserstein-Fisher-Rao
gradient flow.
We prove that, under non-degeneracy assumptions, this method converges at an
exponential rate to the exact mixed Nash equilibrium from any initialization
satisfying a natural notion of closeness to optimality. We illustrate our
results with numerical experiments and discuss applications to max-margin and
distributionally-robust classification using two-layer neural networks, where
our method has a natural interpretation as a simultaneous training of the
network's weights and of the adversarial distribution.
Related papers
- A Unified Theory of Stochastic Proximal Point Methods without Smoothness [52.30944052987393]
Proximal point methods have attracted considerable interest owing to their numerical stability and robustness against imperfect tuning.
This paper presents a comprehensive analysis of a broad range of variations of the proximal point method (SPPM)
arXiv Detail & Related papers (2024-05-24T21:09:19Z) - Collaborative Heterogeneous Causal Inference Beyond Meta-analysis [68.4474531911361]
We propose a collaborative inverse propensity score estimator for causal inference with heterogeneous data.
Our method shows significant improvements over the methods based on meta-analysis when heterogeneity increases.
arXiv Detail & Related papers (2024-04-24T09:04:36Z) - PAPAL: A Provable PArticle-based Primal-Dual ALgorithm for Mixed Nash
Equilibrium [62.51015395213579]
We consider the non-AL equilibrium nonconptotic objective function in two-player zero-sum continuous games.
The proposed algorithm employs the movements of particles to represent the updates of random strategies for the $ilon$-mixed Nash equilibrium.
arXiv Detail & Related papers (2023-03-02T05:08:15Z) - Finding mixed-strategy equilibria of continuous-action games without
gradients using randomized policy networks [83.28949556413717]
We study the problem of computing an approximate Nash equilibrium of continuous-action game without access to gradients.
We model players' strategies using artificial neural networks.
This paper is the first to solve general continuous-action games with unrestricted mixed strategies and without any gradient information.
arXiv Detail & Related papers (2022-11-29T05:16:41Z) - Sampling with Mollified Interaction Energy Descent [57.00583139477843]
We present a new optimization-based method for sampling called mollified interaction energy descent (MIED)
MIED minimizes a new class of energies on probability measures called mollified interaction energies (MIEs)
We show experimentally that for unconstrained sampling problems our algorithm performs on par with existing particle-based algorithms like SVGD.
arXiv Detail & Related papers (2022-10-24T16:54:18Z) - A note on large deviations for interacting particle dynamics for finding
mixed equilibria in zero-sum games [0.0]
Finding equilibria points in continuous minimax games has become a key problem within machine learning.
Recent developments have shifted from pure equilibria to focusing on mixed equilibria points.
We show that the sequence of empirical measures of the particle system satisfies a large deviation principle as the number of particles grows to infinity.
arXiv Detail & Related papers (2022-06-30T10:29:21Z) - Conjugate Gradient Method for Generative Adversarial Networks [0.0]
It is not feasible to calculate the Jensen-Shannon divergence of the density function of the data and the density function of the model of deep neural networks.
Generative adversarial networks (GANs) can be used to formulate this problem as a discriminative problem with two models, a generator and a discriminator.
We propose to apply the conjugate gradient method to solve the local Nash equilibrium problem in GANs.
arXiv Detail & Related papers (2022-03-28T04:44:45Z) - A blob method method for inhomogeneous diffusion with applications to
multi-agent control and sampling [0.6562256987706128]
We develop a deterministic particle method for the weighted porous medium equation (WPME) and prove its convergence on bounded time intervals.
Our method has natural applications to multi-agent coverage algorithms and sampling probability measures.
arXiv Detail & Related papers (2022-02-25T19:49:05Z) - Provably convergent quasistatic dynamics for mean-field two-player
zero-sum games [10.39511271647025]
We consider a quasistatic Wasserstein gradient flow dynamics in which one probability distribution follows the Wasserstein gradient flow, while the other one is always at the equilibrium.
Inspired by the continuous dynamics of probability distributions, we derive a quasistatic Langevin gradient descent method with inner-outer iterations.
arXiv Detail & Related papers (2022-02-15T20:19:42Z) - On the Convergence of Stochastic Extragradient for Bilinear Games with
Restarted Iteration Averaging [96.13485146617322]
We present an analysis of the ExtraGradient (SEG) method with constant step size, and present variations of the method that yield favorable convergence.
We prove that when augmented with averaging, SEG provably converges to the Nash equilibrium, and such a rate is provably accelerated by incorporating a scheduled restarting procedure.
arXiv Detail & Related papers (2021-06-30T17:51:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.