Finite-Time Last-Iterate Convergence for Multi-Agent Learning in Games
- URL: http://arxiv.org/abs/2002.09806v5
- Date: Sat, 17 Jul 2021 06:09:20 GMT
- Title: Finite-Time Last-Iterate Convergence for Multi-Agent Learning in Games
- Authors: Tianyi Lin, Zhengyuan Zhou, Panayotis Mertikopoulos and Michael I.
Jordan
- Abstract summary: We characterize the finite-time last-iterate convergence rate for joint OGD learning on $lambda$-cocoercive games.
We show, via a novel double-stopping time technique, that this adaptive algorithm achieves same finite-time last-iterate convergence rate as non-adaptive counterpart.
- Score: 116.0771177871705
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we consider multi-agent learning via online gradient descent
in a class of games called $\lambda$-cocoercive games, a fairly broad class of
games that admits many Nash equilibria and that properly includes unconstrained
strongly monotone games. We characterize the finite-time last-iterate
convergence rate for joint OGD learning on $\lambda$-cocoercive games; further,
building on this result, we develop a fully adaptive OGD learning algorithm
that does not require any knowledge of problem parameter (e.g. cocoercive
constant $\lambda$) and show, via a novel double-stopping time technique, that
this adaptive algorithm achieves same finite-time last-iterate convergence rate
as non-adaptive counterpart. Subsequently, we extend OGD learning to the noisy
gradient feedback case and establish last-iterate convergence results -- first
qualitative almost sure convergence, then quantitative finite-time convergence
rates -- all under non-decreasing step-sizes. To our knowledge, we provide the
first set of results that fill in several gaps of the existing multi-agent
online learning literature, where three aspects -- finite-time convergence
rates, non-decreasing step-sizes, and fully adaptive algorithms have been
unexplored before.
Related papers
- Adaptive, Doubly Optimal No-Regret Learning in Strongly Monotone and Exp-Concave Games with Gradient Feedback [75.29048190099523]
Online gradient descent (OGD) is well known to be doubly optimal under strong convexity or monotonicity assumptions.
In this paper, we design a fully adaptive OGD algorithm, textsfAdaOGD, that does not require a priori knowledge of these parameters.
arXiv Detail & Related papers (2023-10-21T18:38:13Z) - Smoothing ADMM for Sparse-Penalized Quantile Regression with Non-Convex
Penalties [8.294148737585543]
This paper investigates concave and clipped quantile regression in the presence of nonsecondary absolute and non-smooth convergence penalties.
We introduce a novel-loop ADM algorithm with an increasing penalty multiplier, named SIAD, specifically for sparse regression.
arXiv Detail & Related papers (2023-09-04T21:48:51Z) - On the Convergence of No-Regret Learning Dynamics in Time-Varying Games [89.96815099996132]
We characterize the convergence of optimistic gradient descent (OGD) in time-varying games.
Our framework yields sharp convergence bounds for the equilibrium gap of OGD in zero-sum games.
We also provide new insights on dynamic regret guarantees in static games.
arXiv Detail & Related papers (2023-01-26T17:25:45Z) - Asynchronous Gradient Play in Zero-Sum Multi-agent Games [25.690033495071923]
We study asynchronous gradient plays in zero-sum polymatrix games under delayed feedbacks.
To the best of our knowledge, this work is the first that aims to understand asynchronous gradient play in zero-sum polymatrix games.
arXiv Detail & Related papers (2022-11-16T15:37:23Z) - A unified stochastic approximation framework for learning in games [82.74514886461257]
We develop a flexible approximation framework for analyzing the long-run behavior of learning in games (both continuous and finite)
The proposed analysis template incorporates a wide array of popular learning algorithms, including gradient-based methods, exponential/multiplicative weights for learning in finite games, optimistic and bandit variants of the above, etc.
arXiv Detail & Related papers (2022-06-08T14:30:38Z) - Last-iterate Convergence in Extensive-Form Games [49.31256241275577]
We study last-iterate convergence of optimistic algorithms in sequential games.
We show that all of these algorithms enjoy last-iterate convergence, with some of them even converging exponentially fast.
arXiv Detail & Related papers (2021-06-27T22:02:26Z) - Finite-Time Analysis for Double Q-learning [50.50058000948908]
We provide the first non-asymptotic, finite-time analysis for double Q-learning.
We show that both synchronous and asynchronous double Q-learning are guaranteed to converge to an $epsilon$-accurate neighborhood of the global optimum.
arXiv Detail & Related papers (2020-09-29T18:48:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.