Exponential Convergence of Gradient Methods in Concave Network Zero-sum
Games
- URL: http://arxiv.org/abs/2007.05477v1
- Date: Fri, 10 Jul 2020 16:56:56 GMT
- Title: Exponential Convergence of Gradient Methods in Concave Network Zero-sum
Games
- Authors: Amit Kadan and Hu Fu
- Abstract summary: We study the computation of Nash equilibrium in concave network zero-sum games (NZSGs)
We show that various game theoretic properties of convex-concave two-player zero-sum games are preserved in this generalization.
- Score: 6.129776019898013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motivated by Generative Adversarial Networks, we study the computation of
Nash equilibrium in concave network zero-sum games (NZSGs), a multiplayer
generalization of two-player zero-sum games first proposed with linear payoffs.
Extending previous results, we show that various game theoretic properties of
convex-concave two-player zero-sum games are preserved in this generalization.
We then generalize last iterate convergence results obtained previously in
two-player zero-sum games. We analyze convergence rates when players update
their strategies using Gradient Ascent, and its variant, Optimistic Gradient
Ascent, showing last iterate convergence in three settings -- when the payoffs
of players are linear, strongly concave and Lipschitz, and strongly concave and
smooth. We provide experimental results that support these theoretical
findings.
Related papers
- On the Convergence of No-Regret Learning Dynamics in Time-Varying Games [89.96815099996132]
We characterize the convergence of optimistic gradient descent (OGD) in time-varying games.
Our framework yields sharp convergence bounds for the equilibrium gap of OGD in zero-sum games.
We also provide new insights on dynamic regret guarantees in static games.
arXiv Detail & Related papers (2023-01-26T17:25:45Z) - Improved Convergence Guarantees for Shallow Neural Networks [91.3755431537592]
We prove convergence of depth 2 neural networks, trained via gradient descent, to a global minimum.
Our model has the following features: regression with quadratic loss function, fully connected feedforward architecture, RelU activations, Gaussian data instances, adversarial labels.
They strongly suggest that, at least in our model, the convergence phenomenon extends well beyond the NTK regime''
arXiv Detail & Related papers (2022-12-05T14:47:52Z) - A Unified Approach to Reinforcement Learning, Quantal Response
Equilibria, and Two-Player Zero-Sum Games [104.3339905200105]
This work studies an algorithm, which we call magnetic mirror descent, that is inspired by mirror descent and the non-Euclidean proximal gradient algorithm.
Our contribution is demonstrating the virtues of magnetic mirror descent as both an equilibrium solver and as an approach to reinforcement learning in two-player zero-sum games.
arXiv Detail & Related papers (2022-06-12T19:49:14Z) - Towards convergence to Nash equilibria in two-team zero-sum games [17.4461045395989]
Two-team zero-sum games are defined as multi-player games where players are split into two competing sets of agents.
We focus on the solution concept of Nash equilibria (NE)
We show that computing NE for this class of games is $textithard$ for the complexity class $mathrm$.
arXiv Detail & Related papers (2021-11-07T21:15:35Z) - Last-iterate Convergence in Extensive-Form Games [49.31256241275577]
We study last-iterate convergence of optimistic algorithms in sequential games.
We show that all of these algorithms enjoy last-iterate convergence, with some of them even converging exponentially fast.
arXiv Detail & Related papers (2021-06-27T22:02:26Z) - Better Regularization for Sequential Decision Spaces: Fast Convergence
Rates for Nash, Correlated, and Team Equilibria [121.36609493711292]
We study the application of iterative first-order methods to the problem of computing equilibria of large-scale two-player extensive-form games.
By instantiating first-order methods with our regularizers, we develop the first accelerated first-order methods for computing correlated equilibra and ex-ante coordinated team equilibria.
arXiv Detail & Related papers (2021-05-27T06:10:24Z) - Complex Momentum for Learning in Games [42.081050296353574]
We generalize gradient descent with momentum for learning in differentiable games to have complex-valued momentum.
We empirically demonstrate that complex-valued momentum can improve convergence in games - like generative adversarial networks.
We also show a practical generalization to a complex-valued Adam variant, which we use to train BigGAN to better scores on CIFAR-10.
arXiv Detail & Related papers (2021-02-16T19:55:27Z) - Solving Min-Max Optimization with Hidden Structure via Gradient Descent
Ascent [40.99558326586079]
A class of non-concave zero-sum games functions hidden zero-sum games.
We prove convergence to the von-Neumann equilibrium of the "hidden" convex-concave game.
Our guarantees are non-local, which as far as we know is a first-of-its-kind type of result in non-concave games.
arXiv Detail & Related papers (2021-01-13T18:13:49Z) - From Poincar\'e Recurrence to Convergence in Imperfect Information
Games: Finding Equilibrium via Regularization [49.368421783733815]
We show how adapting the reward can give strong convergence guarantees in monotone games.
We also show how this reward adaptation technique can be leveraged to build algorithms that converge exactly to the Nash equilibrium.
arXiv Detail & Related papers (2020-02-19T21:36:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.