No-Regret Learning in Games with Noisy Feedback: Faster Rates and
Adaptivity via Learning Rate Separation
- URL: http://arxiv.org/abs/2206.06015v1
- Date: Mon, 13 Jun 2022 10:13:51 GMT
- Title: No-Regret Learning in Games with Noisy Feedback: Faster Rates and
Adaptivity via Learning Rate Separation
- Authors: Yu-Guan Hsieh, Kimon Antonakopoulos, Volkan Cevher, Panayotis
Mertikopoulos
- Abstract summary: We study the problem of regret when the learner is involved in a continuous game with other optimizing agents.
In this case, if all players follow a no-regret algorithm, it is possible to achieve significantly lower regret relative to fully adversarial environments.
We propose a fully adaptive method that smoothly interpolates between worst- and best-case regret guarantees.
- Score: 76.61911795703062
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We examine the problem of regret minimization when the learner is involved in
a continuous game with other optimizing agents: in this case, if all players
follow a no-regret algorithm, it is possible to achieve significantly lower
regret relative to fully adversarial environments. We study this problem in the
context of variationally stable games (a class of continuous games which
includes all convex-concave and monotone games), and when the players only have
access to noisy estimates of their individual payoff gradients. If the noise is
additive, the game-theoretic and purely adversarial settings enjoy similar
regret guarantees; however, if the noise is multiplicative, we show that the
learners can, in fact, achieve constant regret. We achieve this faster rate via
an optimistic gradient scheme with learning rate separation -- that is, the
method's extrapolation and update steps are tuned to different schedules,
depending on the noise profile. Subsequently, to eliminate the need for
delicate hyperparameter tuning, we propose a fully adaptive method that
smoothly interpolates between worst- and best-case regret guarantees.
Related papers
- Online Convex Optimisation: The Optimal Switching Regret for all Segmentations Simultaneously [8.850922234275636]
A switching regret is defined relative to any segmentation of the trial sequence, and is equal to the sum of the static regrets of each segment.
Our algorithm for doing so is very efficient: having a space and per-trial time complexity that is logarithmic in the time-horizon.
arXiv Detail & Related papers (2024-05-31T14:16:52Z) - Advancing Unsupervised Low-light Image Enhancement: Noise Estimation, Illumination Interpolation, and Self-Regulation [55.07472635587852]
Low-Light Image Enhancement (LLIE) techniques have made notable advancements in preserving image details and enhancing contrast.
These approaches encounter persistent challenges in efficiently mitigating dynamic noise and accommodating diverse low-light scenarios.
We first propose a method for estimating the noise level in low light images in a quick and accurate way.
We then devise a Learnable Illumination Interpolator (LII) to satisfy general constraints between illumination and input.
arXiv Detail & Related papers (2023-05-17T13:56:48Z) - On the Convergence of No-Regret Learning Dynamics in Time-Varying Games [89.96815099996132]
We characterize the convergence of optimistic gradient descent (OGD) in time-varying games.
Our framework yields sharp convergence bounds for the equilibrium gap of OGD in zero-sum games.
We also provide new insights on dynamic regret guarantees in static games.
arXiv Detail & Related papers (2023-01-26T17:25:45Z) - Finding mixed-strategy equilibria of continuous-action games without
gradients using randomized policy networks [83.28949556413717]
We study the problem of computing an approximate Nash equilibrium of continuous-action game without access to gradients.
We model players' strategies using artificial neural networks.
This paper is the first to solve general continuous-action games with unrestricted mixed strategies and without any gradient information.
arXiv Detail & Related papers (2022-11-29T05:16:41Z) - Regret Minimization and Convergence to Equilibria in General-sum Markov
Games [57.568118148036376]
We present the first algorithm for learning in general-sum Markov games that provides sublinear regret guarantees when executed by all agents.
Our algorithm is decentralized, computationally efficient, and does not require any communication between agents.
arXiv Detail & Related papers (2022-07-28T16:27:59Z) - Efficient Adaptive Regret Minimization [35.121567896321885]
In online convex optimization the player aims to minimize her regret against a fixed comparator over the entire repeated game.
Existing adaptive regret algorithms suffer from a computational penalty - typically on the order of a multiplicative factor that grows logarithmically in the number of game iterations.
In this paper we show how to reduce this computational penalty to be doubly logarithmic in the number of game iterations, and with minimal degradation to the optimal attainable adaptive regret bounds.
arXiv Detail & Related papers (2022-07-01T19:43:11Z) - Adaptive Learning in Continuous Games: Optimal Regret Bounds and
Convergence to Nash Equilibrium [33.9962699667578]
No-regret algorithms are not created equal in terms of game-theoretic guarantees.
We propose a range of no-regret policies based on optimistic mirror descent.
arXiv Detail & Related papers (2021-04-26T17:52:29Z) - Model-Free Online Learning in Unknown Sequential Decision Making
Problems and Games [114.90723492840499]
In large two-player zero-sum imperfect-information games, modern extensions of counterfactual regret minimization (CFR) are currently the practical state of the art for computing a Nash equilibrium.
We formalize an online learning setting in which the strategy space is not known to the agent.
We give an efficient algorithm that achieves $O(T3/4)$ regret with high probability for that setting, even when the agent faces an adversarial environment.
arXiv Detail & Related papers (2021-03-08T04:03:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.