Exponential Convergence of Gradient Methods in Concave Network Zero-sum
  Games
        - URL: http://arxiv.org/abs/2007.05477v1
 - Date: Fri, 10 Jul 2020 16:56:56 GMT
 - Title: Exponential Convergence of Gradient Methods in Concave Network Zero-sum
  Games
 - Authors: Amit Kadan and Hu Fu
 - Abstract summary: We study the computation of Nash equilibrium in concave network zero-sum games (NZSGs)
We show that various game theoretic properties of convex-concave two-player zero-sum games are preserved in this generalization.
 - Score: 6.129776019898013
 - License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
 - Abstract:   Motivated by Generative Adversarial Networks, we study the computation of
Nash equilibrium in concave network zero-sum games (NZSGs), a multiplayer
generalization of two-player zero-sum games first proposed with linear payoffs.
Extending previous results, we show that various game theoretic properties of
convex-concave two-player zero-sum games are preserved in this generalization.
We then generalize last iterate convergence results obtained previously in
two-player zero-sum games. We analyze convergence rates when players update
their strategies using Gradient Ascent, and its variant, Optimistic Gradient
Ascent, showing last iterate convergence in three settings -- when the payoffs
of players are linear, strongly concave and Lipschitz, and strongly concave and
smooth. We provide experimental results that support these theoretical
findings.
 
       
      
        Related papers
        - From Average-Iterate to Last-Iterate Convergence in Games: A Reduction   and Its Applications [44.95137108337898]
We show that for a large family of games, there exists a simple black-box reduction that transforms the average iterates of an uncoupled learning dynamics into the last iterates of a new uncoupled learning dynamics.<n>We obtain new state-of-the-art last-iterate convergence rates for uncoupled learning dynamics in two-player zero-sum normal-form games.
arXiv  Detail & Related papers  (2025-06-04T00:24:14Z) - Last-Iterate Convergence Properties of Regret-Matching Algorithms in   Games [71.73971094342349]
We study last-iterate convergence properties of algorithms for solving two-player zero-sum games based on Regret$+$ (RM$+$)
Our analysis builds on a new characterization of the geometric structure of the limit points of our algorithms, marking a significant departure from most of the literature on last-iterate convergence.
arXiv  Detail & Related papers  (2023-11-01T17:34:58Z) - On the Convergence of No-Regret Learning Dynamics in Time-Varying Games [89.96815099996132]
We characterize the convergence of optimistic gradient descent (OGD) in time-varying games.
Our framework yields sharp convergence bounds for the equilibrium gap of OGD in zero-sum games.
We also provide new insights on dynamic regret guarantees in static games.
arXiv  Detail & Related papers  (2023-01-26T17:25:45Z) - Improved Convergence Guarantees for Shallow Neural Networks [91.3755431537592]
We prove convergence of depth 2 neural networks, trained via gradient descent, to a global minimum.
Our model has the following features: regression with quadratic loss function, fully connected feedforward architecture, RelU activations, Gaussian data instances, adversarial labels.
They strongly suggest that, at least in our model, the convergence phenomenon extends well beyond the NTK regime''
arXiv  Detail & Related papers  (2022-12-05T14:47:52Z) - A Unified Approach to Reinforcement Learning, Quantal Response
  Equilibria, and Two-Player Zero-Sum Games [104.3339905200105]
This work studies an algorithm, which we call magnetic mirror descent, that is inspired by mirror descent and the non-Euclidean proximal gradient algorithm.
Our contribution is demonstrating the virtues of magnetic mirror descent as both an equilibrium solver and as an approach to reinforcement learning in two-player zero-sum games.
arXiv  Detail & Related papers  (2022-06-12T19:49:14Z) - Towards convergence to Nash equilibria in two-team zero-sum games [17.4461045395989]
Two-team zero-sum games are defined as multi-player games where players are split into two competing sets of agents.
We focus on the solution concept of Nash equilibria (NE)
We show that computing NE for this class of games is $textithard$ for the complexity class $mathrm$.
arXiv  Detail & Related papers  (2021-11-07T21:15:35Z) - Last-iterate Convergence in Extensive-Form Games [49.31256241275577]
We study last-iterate convergence of optimistic algorithms in sequential games.
We show that all of these algorithms enjoy last-iterate convergence, with some of them even converging exponentially fast.
arXiv  Detail & Related papers  (2021-06-27T22:02:26Z) - Better Regularization for Sequential Decision Spaces: Fast Convergence
  Rates for Nash, Correlated, and Team Equilibria [121.36609493711292]
We study the application of iterative first-order methods to the problem of computing equilibria of large-scale two-player extensive-form games.
By instantiating first-order methods with our regularizers, we develop the first accelerated first-order methods for computing correlated equilibra and ex-ante coordinated team equilibria.
arXiv  Detail & Related papers  (2021-05-27T06:10:24Z) - Complex Momentum for Learning in Games [42.081050296353574]
We generalize gradient descent with momentum for learning in differentiable games to have complex-valued momentum.
We empirically demonstrate that complex-valued momentum can improve convergence in games - like generative adversarial networks.
We also show a practical generalization to a complex-valued Adam variant, which we use to train BigGAN to better scores on CIFAR-10.
arXiv  Detail & Related papers  (2021-02-16T19:55:27Z) - Solving Min-Max Optimization with Hidden Structure via Gradient Descent
  Ascent [40.99558326586079]
A class of non-concave zero-sum games functions hidden zero-sum games.
We prove convergence to the von-Neumann equilibrium of the "hidden" convex-concave game.
Our guarantees are non-local, which as far as we know is a first-of-its-kind type of result in non-concave games.
arXiv  Detail & Related papers  (2021-01-13T18:13:49Z) - From Poincar\'e Recurrence to Convergence in Imperfect Information
  Games: Finding Equilibrium via Regularization [49.368421783733815]
We show how adapting the reward can give strong convergence guarantees in monotone games.
We also show how this reward adaptation technique can be leveraged to build algorithms that converge exactly to the Nash equilibrium.
arXiv  Detail & Related papers  (2020-02-19T21:36:58Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.