Newton Optimization on Helmholtz Decomposition for Continuous Games
- URL: http://arxiv.org/abs/2007.07804v3
- Date: Thu, 2 Sep 2021 12:01:18 GMT
- Title: Newton Optimization on Helmholtz Decomposition for Continuous Games
- Authors: Giorgia Ramponi and Marcello Restelli
- Abstract summary: NOHD is a Newton-like algorithm for multi-agent learning problems based on the decomposition of the dynamics of the system.
We show that NOHD is attracted to stable fixed points in general multi-agent systems and repelled by strict saddle ones.
- Score: 47.42898331586512
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many learning problems involve multiple agents optimizing different
interactive functions. In these problems, the standard policy gradient
algorithms fail due to the non-stationarity of the setting and the different
interests of each agent. In fact, algorithms must take into account the complex
dynamics of these systems to guarantee rapid convergence towards a (local) Nash
equilibrium. In this paper, we propose NOHD (Newton Optimization on Helmholtz
Decomposition), a Newton-like algorithm for multi-agent learning problems based
on the decomposition of the dynamics of the system in its irrotational
(Potential) and solenoidal (Hamiltonian) component. This method ensures
quadratic convergence in purely irrotational systems and pure solenoidal
systems. Furthermore, we show that NOHD is attracted to stable fixed points in
general multi-agent systems and repelled by strict saddle ones. Finally, we
empirically compare the NOHD's performance with that of state-of-the-art
algorithms on some bimatrix games and in a continuous Gridworld environment.
Related papers
- Reduced-Space Iteratively Reweighted Second-Order Methods for Nonconvex Sparse Regularization [11.56128809794923]
This paper explores a specific type of non sparsity-promoting regularization problems, namely those involving $ell_p-$ iterations of local property convergence.
arXiv Detail & Related papers (2024-07-24T12:15:59Z) - Independent Learning in Constrained Markov Potential Games [19.083595175045073]
Constrained Markov games offer a formal framework for modeling multi-agent reinforcement learning problems.
We propose an independent policy gradient algorithm for learning approximate constrained Nash equilibria.
arXiv Detail & Related papers (2024-02-27T20:57:35Z) - PAPAL: A Provable PArticle-based Primal-Dual ALgorithm for Mixed Nash
Equilibrium [62.51015395213579]
We consider the non-AL equilibrium nonconptotic objective function in two-player zero-sum continuous games.
The proposed algorithm employs the movements of particles to represent the updates of random strategies for the $ilon$-mixed Nash equilibrium.
arXiv Detail & Related papers (2023-03-02T05:08:15Z) - Regularized Gradient Descent Ascent for Two-Player Zero-Sum Markov Games [16.09467599829253]
We study the problem of finding Nash equilibrium in a two-player zero-sum game.
Our main contribution is to show that under proper choices of the regularization parameter, the gradient descents to the Nash equilibrium of the original unregularized problem.
arXiv Detail & Related papers (2022-05-27T03:24:12Z) - First-Order Algorithms for Nonlinear Generalized Nash Equilibrium
Problems [88.58409977434269]
We consider the problem of computing an equilibrium in a class of nonlinear generalized Nash equilibrium problems (NGNEPs)
Our contribution is to provide two simple first-order algorithmic frameworks based on the quadratic penalty method and the augmented Lagrangian method.
We provide nonasymptotic theoretical guarantees for these algorithms.
arXiv Detail & Related papers (2022-04-07T00:11:05Z) - Message Passing Neural PDE Solvers [60.77761603258397]
We build a neural message passing solver, replacing allally designed components in the graph with backprop-optimized neural function approximators.
We show that neural message passing solvers representationally contain some classical methods, such as finite differences, finite volumes, and WENO schemes.
We validate our method on various fluid-like flow problems, demonstrating fast, stable, and accurate performance across different domain topologies, equation parameters, discretizations, etc., in 1D and 2D.
arXiv Detail & Related papers (2022-02-07T17:47:46Z) - Optimization on manifolds: A symplectic approach [127.54402681305629]
We propose a dissipative extension of Dirac's theory of constrained Hamiltonian systems as a general framework for solving optimization problems.
Our class of (accelerated) algorithms are not only simple and efficient but also applicable to a broad range of contexts.
arXiv Detail & Related papers (2021-07-23T13:43:34Z) - Reinforcement Learning with Fast Stabilization in Linear Dynamical
Systems [91.43582419264763]
We study model-based reinforcement learning (RL) in unknown stabilizable linear dynamical systems.
We propose an algorithm that certifies fast stabilization of the underlying system by effectively exploring the environment.
We show that the proposed algorithm attains $tildemathcalO(sqrtT)$ regret after $T$ time steps of agent-environment interaction.
arXiv Detail & Related papers (2020-07-23T23:06:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.