Game Theoretic Rating in N-player general-sum games with Equilibria
- URL: http://arxiv.org/abs/2210.02205v1
- Date: Wed, 5 Oct 2022 12:33:03 GMT
- Title: Game Theoretic Rating in N-player general-sum games with Equilibria
- Authors: Luke Marris, Marc Lanctot, Ian Gemp, Shayegan Omidshafiei, Stephen
McAleer, Jerome Connor, Karl Tuyls, Thore Graepel
- Abstract summary: We propose novel algorithms suitable for N-player, general-sum rating of strategies in normal-form games according to the payoff rating system.
This enables well-established solution concepts, such as equilibria, to be leveraged to efficiently rate strategies in games with complex strategic interactions.
- Score: 26.166859475522106
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Rating strategies in a game is an important area of research in game theory
and artificial intelligence, and can be applied to any real-world competitive
or cooperative setting. Traditionally, only transitive dependencies between
strategies have been used to rate strategies (e.g. Elo), however recent work
has expanded ratings to utilize game theoretic solutions to better rate
strategies in non-transitive games. This work generalizes these ideas and
proposes novel algorithms suitable for N-player, general-sum rating of
strategies in normal-form games according to the payoff rating system. This
enables well-established solution concepts, such as equilibria, to be leveraged
to efficiently rate strategies in games with complex strategic interactions,
which arise in multiagent training and real-world interactions between many
agents. We empirically validate our methods on real world normal-form data
(Premier League) and multiagent reinforcement learning agent evaluation.
Related papers
- Toward Optimal LLM Alignments Using Two-Player Games [86.39338084862324]
In this paper, we investigate alignment through the lens of two-agent games, involving iterative interactions between an adversarial and a defensive agent.
We theoretically demonstrate that this iterative reinforcement learning optimization converges to a Nash Equilibrium for the game induced by the agents.
Experimental results in safety scenarios demonstrate that learning in such a competitive environment not only fully trains agents but also leads to policies with enhanced generalization capabilities for both adversarial and defensive agents.
arXiv Detail & Related papers (2024-06-16T15:24:50Z) - ALYMPICS: LLM Agents Meet Game Theory -- Exploring Strategic
Decision-Making with AI Agents [77.34720446306419]
Alympics is a systematic simulation framework utilizing Large Language Model (LLM) agents for game theory research.
Alympics creates a versatile platform for studying complex game theory problems.
arXiv Detail & Related papers (2023-11-06T16:03:46Z) - All by Myself: Learning Individualized Competitive Behaviour with a
Contrastive Reinforcement Learning optimization [57.615269148301515]
In a competitive game scenario, a set of agents have to learn decisions that maximize their goals and minimize their adversaries' goals at the same time.
We propose a novel model composed of three neural layers that learn a representation of a competitive game, learn how to map the strategy of specific opponents, and how to disrupt them.
Our experiments demonstrate that our model achieves better performance when playing against offline, online, and competitive-specific models, in particular when playing against the same opponent multiple times.
arXiv Detail & Related papers (2023-10-02T08:11:07Z) - Co-Learning Empirical Games and World Models [23.800790782022222]
Empirical games drive world models toward a broader consideration of possible game dynamics.
World models guide empirical games to efficiently discover new strategies through planning.
A new algorithm, Dyna-PSRO, co-learns an empirical game and a world model.
arXiv Detail & Related papers (2023-05-23T16:37:21Z) - Opponent Modeling in Multiplayer Imperfect-Information Games [1.024113475677323]
We present an approach for opponent modeling in multiplayer imperfect-information games.
We run experiments against a variety of real opponents and exact Nash equilibrium strategies in three-player Kuhn poker.
Our algorithm significantly outperforms all of the agents, including the exact Nash equilibrium strategies.
arXiv Detail & Related papers (2022-12-12T16:48:53Z) - Finding mixed-strategy equilibria of continuous-action games without
gradients using randomized policy networks [83.28949556413717]
We study the problem of computing an approximate Nash equilibrium of continuous-action game without access to gradients.
We model players' strategies using artificial neural networks.
This paper is the first to solve general continuous-action games with unrestricted mixed strategies and without any gradient information.
arXiv Detail & Related papers (2022-11-29T05:16:41Z) - Discovering Diverse Multi-Agent Strategic Behavior via Reward
Randomization [42.33734089361143]
We propose a technique for discovering diverse strategic policies in complex multi-agent games.
We derive a new algorithm, Reward-Randomized Policy Gradient (RPG)
RPG is able to discover multiple distinctive human-interpretable strategies in challenging temporal trust dilemmas.
arXiv Detail & Related papers (2021-03-08T06:26:55Z) - On the Impossibility of Convergence of Mixed Strategies with No Regret
Learning [10.515544361834241]
We study convergence properties of the mixed strategies that result from a general class of optimal no regret learning strategies.
We consider the class of strategies whose information set at each step is the empirical average of the opponent's realized play.
arXiv Detail & Related papers (2020-12-03T18:02:40Z) - The Design Of "Stratega": A General Strategy Games Framework [62.997667081978825]
Stratega is a framework for creating turn-based and real-time strategy games.
The framework has been built with a focus on statistical forward planning (SFP) agents.
We hope that the development of this framework and its respective agents helps to better understand the complex decision-making process in strategy games.
arXiv Detail & Related papers (2020-09-11T20:02:00Z) - Learning to Play No-Press Diplomacy with Best Response Policy Iteration [31.367850729299665]
We apply deep reinforcement learning methods to Diplomacy, a 7-player board game.
We show that our agents convincingly outperform the previous state-of-the-art, and game theoretic equilibrium analysis shows that the new process yields consistent improvements.
arXiv Detail & Related papers (2020-06-08T14:33:31Z) - Efficient exploration of zero-sum stochastic games [83.28949556413717]
We investigate the increasingly important and common game-solving setting where we do not have an explicit description of the game but only oracle access to it through gameplay.
During a limited-duration learning phase, the algorithm can control the actions of both players in order to try to learn the game and how to play it well.
Our motivation is to quickly learn strategies that have low exploitability in situations where evaluating the payoffs of a queried strategy profile is costly.
arXiv Detail & Related papers (2020-02-24T20:30:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.