Learning to Beat ByteRL: Exploitability of Collectible Card Game Agents
- URL: http://arxiv.org/abs/2404.16689v1
- Date: Thu, 25 Apr 2024 15:48:40 GMT
- Title: Learning to Beat ByteRL: Exploitability of Collectible Card Game Agents
- Authors: Radovan Haluska, Martin Schmid,
- Abstract summary: We present preliminary analysis results of ByteRL, the state-of-the-art agent in Legends of Code and Magic and Hearthstone.
Although ByteRL beat a top-10 Hearthstone player from China, we show that its play in Legends of Code and Magic is highly exploitable.
- Score: 3.5036467860577307
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While Poker, as a family of games, has been studied extensively in the last decades, collectible card games have seen relatively little attention. Only recently have we seen an agent that can compete with professional human players in Hearthstone, one of the most popular collectible card games. Although artificial agents must be able to work with imperfect information in both of these genres, collectible card games pose another set of distinct challenges. Unlike in many poker variants, agents must deal with state space so vast that even enumerating all states consistent with the agent's beliefs is intractable, rendering the current search methods unusable and requiring the agents to opt for other techniques. In this paper, we investigate the strength of such techniques for this class of games. Namely, we present preliminary analysis results of ByteRL, the state-of-the-art agent in Legends of Code and Magic and Hearthstone. Although ByteRL beat a top-10 Hearthstone player from China, we show that its play in Legends of Code and Magic is highly exploitable.
Related papers
- Optimizing Hearthstone Agents using an Evolutionary Algorithm [0.0]
This paper proposes the use of evolutionary algorithms (EAs) to develop agents who play a card game, Hearthstone.
Agents feature self-learning by means of a competitive coevolutionary training approach.
One of the agents developed through the proposed approach was runner-up (best 6%) in an international Hearthstone Artificial Intelligence (AI) competition.
arXiv Detail & Related papers (2024-10-25T16:49:11Z) - A Survey on Game Theory Optimal Poker [0.0]
No non-trivial imperfect information game has been solved to date.
This makes poker a great test bed for Artificial Intelligence research.
We discuss the intricacies of abstraction techniques, betting models, and specific strategies employed by successful poker bots.
arXiv Detail & Related papers (2024-01-02T04:19:25Z) - Mastering Strategy Card Game (Hearthstone) with Improved Techniques [8.399453146308502]
Strategy card game is demanding on the intelligent game-play and can be an ideal test-bench for AI.
Previous work combines an end-to-end policy function and an optimistic smooth fictitious play.
In this work, we apply such algorithms to Hearthstone, a famous commercial game that is more complicated in game rules and mechanisms.
arXiv Detail & Related papers (2023-03-09T11:52:52Z) - Are AlphaZero-like Agents Robust to Adversarial Perturbations? [73.13944217915089]
AlphaZero (AZ) has demonstrated that neural-network-based Go AIs can surpass human performance by a large margin.
We ask whether adversarial states exist for Go AIs that may lead them to play surprisingly wrong actions.
We develop the first adversarial attack on Go AIs that can efficiently search for adversarial states by strategically reducing the search space.
arXiv Detail & Related papers (2022-11-07T18:43:25Z) - Mastering the Game of No-Press Diplomacy via Human-Regularized
Reinforcement Learning and Planning [95.78031053296513]
No-press Diplomacy is a complex strategy game involving both cooperation and competition.
We introduce a planning algorithm we call DiL-piKL that regularizes a reward-maximizing policy toward a human imitation-learned policy.
We show that DiL-piKL can be extended into a self-play reinforcement learning algorithm we call RL-DiL-piKL.
arXiv Detail & Related papers (2022-10-11T14:47:35Z) - A Technique to Create Weaker Abstract Board Game Agents via
Reinforcement Learning [0.0]
Board games need at least one other player to play.
We created AI agents to play against us when an opponent is missing.
In this work, we describe how to create weaker AI agents that play board games.
arXiv Detail & Related papers (2022-09-01T20:13:20Z) - Mastering the Game of Stratego with Model-Free Multiagent Reinforcement
Learning [86.37438204416435]
Stratego is one of the few iconic board games that Artificial Intelligence (AI) has not yet mastered.
Decisions in Stratego are made over a large number of discrete actions with no obvious link between action and outcome.
DeepNash beats existing state-of-the-art AI methods in Stratego and achieved a yearly (2022) and all-time top-3 rank on the Gravon games platform.
arXiv Detail & Related papers (2022-06-30T15:53:19Z) - Optimisation of MCTS Player for The Lord of the Rings: The Card Game [0.0]
The article presents research on the use of Monte-Carlo Tree Search (MCTS) methods to create an artificial player for the popular card game "The Lord of the Rings"
arXiv Detail & Related papers (2021-09-24T14:42:32Z) - Generating Diverse and Competitive Play-Styles for Strategy Games [58.896302717975445]
We propose Portfolio Monte Carlo Tree Search with Progressive Unpruning for playing a turn-based strategy game (Tribes)
We show how it can be parameterized so a quality-diversity algorithm (MAP-Elites) is used to achieve different play-styles while keeping a competitive level of play.
Our results show that this algorithm is capable of achieving these goals even for an extensive collection of game levels beyond those used for training.
arXiv Detail & Related papers (2021-04-17T20:33:24Z) - Multi-Agent Collaboration via Reward Attribution Decomposition [75.36911959491228]
We propose Collaborative Q-learning (CollaQ) that achieves state-of-the-art performance in the StarCraft multi-agent challenge.
CollaQ is evaluated on various StarCraft Attribution maps and shows that it outperforms existing state-of-the-art techniques.
arXiv Detail & Related papers (2020-10-16T17:42:11Z) - Learning from Learners: Adapting Reinforcement Learning Agents to be
Competitive in a Card Game [71.24825724518847]
We present a study on how popular reinforcement learning algorithms can be adapted to learn and to play a real-world implementation of a competitive multiplayer card game.
We propose specific training and validation routines for the learning agents, in order to evaluate how the agents learn to be competitive and explain how they adapt to each others' playing style.
arXiv Detail & Related papers (2020-04-08T14:11:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.