DecisionHoldem: Safe Depth-Limited Solving With Diverse Opponents for Imperfect-Information Games
- URL: http://arxiv.org/abs/2201.11580v2
- Date: Tue, 28 May 2024 05:04:52 GMT
- Title: DecisionHoldem: Safe Depth-Limited Solving With Diverse Opponents for Imperfect-Information Games
- Authors: Qibin Zhou, Dongdong Bai, Junge Zhang, Fuqing Duan, Kaiqi Huang,
- Abstract summary: DecisionHoldem is a high-level AI for heads-up no-limit Texas hold'em with safe depth-limited subgame solving.
We release the source codes and tools of DecisionHoldem to promote AI development in imperfect-information games.
- Score: 31.26667266662521
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An imperfect-information game is a type of game with asymmetric information. It is more common in life than perfect-information game. Artificial intelligence (AI) in imperfect-information games, such like poker, has made considerable progress and success in recent years. The great success of superhuman poker AI, such as Libratus and Deepstack, attracts researchers to pay attention to poker research. However, the lack of open-source code limits the development of Texas hold'em AI to some extent. This article introduces DecisionHoldem, a high-level AI for heads-up no-limit Texas hold'em with safe depth-limited subgame solving by considering possible ranges of opponent's private hands to reduce the exploitability of the strategy. Experimental results show that DecisionHoldem defeats the strongest openly available agent in heads-up no-limit Texas hold'em poker, namely Slumbot, and a high-level reproduction of Deepstack, viz, Openstack, by more than 730 mbb/h (one-thousandth big blind per round) and 700 mbb/h. Moreover, we release the source codes and tools of DecisionHoldem to promote AI development in imperfect-information games.
Related papers
- A Survey on Game Theory Optimal Poker [0.0]
No non-trivial imperfect information game has been solved to date.
This makes poker a great test bed for Artificial Intelligence research.
We discuss the intricacies of abstraction techniques, betting models, and specific strategies employed by successful poker bots.
arXiv Detail & Related papers (2024-01-02T04:19:25Z) - DanZero+: Dominating the GuanDan Game through Reinforcement Learning [95.90682269990705]
We develop an AI program for an exceptionally complex and popular card game called GuanDan.
We first put forward an AI program named DanZero for this game.
In order to further enhance the AI's capabilities, we apply policy-based reinforcement learning algorithm to GuanDan.
arXiv Detail & Related papers (2023-12-05T08:07:32Z) - Are AlphaZero-like Agents Robust to Adversarial Perturbations? [73.13944217915089]
AlphaZero (AZ) has demonstrated that neural-network-based Go AIs can surpass human performance by a large margin.
We ask whether adversarial states exist for Go AIs that may lead them to play surprisingly wrong actions.
We develop the first adversarial attack on Go AIs that can efficiently search for adversarial states by strategically reducing the search space.
arXiv Detail & Related papers (2022-11-07T18:43:25Z) - DanZero: Mastering GuanDan Game with Reinforcement Learning [121.93690719186412]
Card game AI has always been a hot topic in the research of artificial intelligence.
In this paper, we are devoted to developing an AI program for a more complex card game, GuanDan.
We propose the first AI program DanZero for GuanDan using reinforcement learning technique.
arXiv Detail & Related papers (2022-10-31T06:29:08Z) - Mastering the Game of Stratego with Model-Free Multiagent Reinforcement
Learning [86.37438204416435]
Stratego is one of the few iconic board games that Artificial Intelligence (AI) has not yet mastered.
Decisions in Stratego are made over a large number of discrete actions with no obvious link between action and outcome.
DeepNash beats existing state-of-the-art AI methods in Stratego and achieved a yearly (2022) and all-time top-3 rank on the Gravon games platform.
arXiv Detail & Related papers (2022-06-30T15:53:19Z) - Using Graph Convolutional Networks and TD($\lambda$) to play the game of
Risk [0.0]
Risk is a 6 player game with significant randomness and a large game-tree complexity.
Previous AIs focus on creating high-level handcrafted features determine agent decision making.
I create D.A.D, A Risk agent using temporal difference reinforcement learning to train a Deep Neural Network.
arXiv Detail & Related papers (2020-09-10T18:47:08Z) - Suphx: Mastering Mahjong with Deep Reinforcement Learning [114.68233321904623]
We design an AI for Mahjong, named Suphx, based on deep reinforcement learning with some newly introduced techniques.
Suphx has demonstrated stronger performance than most top human players in terms of stable rank.
This is the first time that a computer program outperforms most top human players in Mahjong.
arXiv Detail & Related papers (2020-03-30T16:18:16Z) - From Chess and Atari to StarCraft and Beyond: How Game AI is Driving the
World of AI [10.80914659291096]
Game AI has established itself as a research area for developing and testing the most advanced forms of AI algorithms.
Advances in Game AI are starting to be extended to areas outside of games, such as robotics or the synthesis of chemicals.
arXiv Detail & Related papers (2020-02-24T18:28:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.