Competing in a Complex Hidden Role Game with Information Set Monte Carlo
Tree Search
- URL: http://arxiv.org/abs/2005.07156v1
- Date: Thu, 14 May 2020 17:21:10 GMT
- Title: Competing in a Complex Hidden Role Game with Information Set Monte Carlo
Tree Search
- Authors: Jack Reinhardt
- Abstract summary: The Information Set Monte Carlo Tree Search (ISMCTS) family of algorithms outperforms previous algorithms using Monte Carlo methods in imperfect information games.
This paper is applied to Secret Hitler, a popular social deduction board game that combines traditional hidden role mechanics with the randomness of a card deck.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Advances in intelligent game playing agents have led to successes in perfect
information games like Go and imperfect information games like Poker. The
Information Set Monte Carlo Tree Search (ISMCTS) family of algorithms
outperforms previous algorithms using Monte Carlo methods in imperfect
information games. In this paper, Single Observer Information Set Monte Carlo
Tree Search (SO-ISMCTS) is applied to Secret Hitler, a popular social deduction
board game that combines traditional hidden role mechanics with the randomness
of a card deck. This combination leads to a more complex information model than
the hidden role and card deck mechanics alone. It is shown in 10108 simulated
games that SO-ISMCTS plays as well as simpler rule based agents, and
demonstrates the potential of ISMCTS algorithms in complicated information set
domains.
Related papers
- Improve Value Estimation of Q Function and Reshape Reward with Monte Carlo Tree Search [0.4450107621124637]
Reinforcement learning has achieved remarkable success in perfect information games such as Go and Atari.
Research in reinforcement learning for imperfect information games has been relatively limited due to the more complex game structures and randomness.
In this paper, we focus on Uno, an imperfect information game, and aim to address these problems by reducing Q value overestimation and reshaping reward function.
arXiv Detail & Related papers (2024-10-15T14:31:54Z) - Transformer Based Planning in the Observation Space with Applications to Trick Taking Card Games [14.864985236886314]
We present Generative Observation Monte Carlo Tree Search (GO-MCTS)
GO-MCTS searches within the observation space and advances the search using a model that depends solely on the agent's observations.
The efficacy of GO-MCTS is demonstrated in various games of imperfect information, such as Hearts, Skat, and "The Crew: The Quest for Planet Nine"
arXiv Detail & Related papers (2024-04-19T19:41:00Z) - Monte-Carlo Tree Search for Multi-Agent Pathfinding: Preliminary Results [60.4817465598352]
We introduce an original variant of Monte-Carlo Tree Search (MCTS) tailored to multi-agent pathfinding.
Specifically, we use individual paths to assist the agents with the the goal-reaching behavior.
We also use a dedicated decomposition technique to reduce the branching factor of the tree search procedure.
arXiv Detail & Related papers (2023-07-25T12:33:53Z) - Regret Matching+: (In)Stability and Fast Convergence in Games [68.13214224119024]
We show that RM+ and its predictive version can be unstable, which might cause other players to suffer large regret.
We show that these fixes are sufficient to get $O(T1/4)$ individual regret and $O(1)$ social regret in normal-form games via RM+ with predictions.
arXiv Detail & Related papers (2023-05-24T04:26:21Z) - The Update-Equivalence Framework for Decision-Time Planning [78.44953498421854]
We introduce an alternative framework for decision-time planning that is not based on solving subgames, but rather on update equivalence.
We derive a provably sound search algorithm for fully cooperative games based on mirror descent and a search algorithm for adversarial games based on magnetic mirror descent.
arXiv Detail & Related papers (2023-04-25T20:28:55Z) - MCTS Based Agents for Multistage Single-Player Card Game [0.0]
The article presents the use of Monte Carlo Tree Search algorithms for the card game Lord of the Rings.
The main challenge was the complexity of the game mechanics, in which each round consists of 5 decision stages and 2 random stages.
To test various decision-making algorithms, a game simulator has been implemented.
arXiv Detail & Related papers (2021-09-24T10:56:54Z) - Learning to Play Imperfect-Information Games by Imitating an Oracle
Planner [77.67437357688316]
We consider learning to play multiplayer imperfect-information games with simultaneous moves and large state-action spaces.
Our approach is based on model-based planning.
We show that the planner is able to discover efficient playing strategies in the games of Clash Royale and Pommerman.
arXiv Detail & Related papers (2020-12-22T17:29:57Z) - Using Tabu Search Algorithm for Map Generation in the Terra Mystica
Tabletop Game [60.71662712899962]
Tabu Search (TS) metaheuristic improves simple local search algorithms by enabling the algorithm to escape local optima points.
This paper investigates the performance of TS and considers the effects of the size of the Tabu list and the size of the neighbourhood for a procedural content generation.
arXiv Detail & Related papers (2020-06-04T09:15:46Z) - Single-Agent Optimization Through Policy Iteration Using Monte-Carlo
Tree Search [8.22379888383833]
Combination of Monte-Carlo Tree Search (MCTS) and deep reinforcement learning is state-of-the-art in two-player perfect-information games.
We describe a search algorithm that uses a variant of MCTS which we enhanced by 1) a novel action value normalization mechanism for games with potentially unbounded rewards, 2) defining a virtual loss function that enables effective search parallelization, and 3) a policy network, trained by generations of self-play, to guide the search.
arXiv Detail & Related papers (2020-05-22T18:02:36Z) - From Poincar\'e Recurrence to Convergence in Imperfect Information
Games: Finding Equilibrium via Regularization [49.368421783733815]
We show how adapting the reward can give strong convergence guarantees in monotone games.
We also show how this reward adaptation technique can be leveraged to build algorithms that converge exactly to the Nash equilibrium.
arXiv Detail & Related papers (2020-02-19T21:36:58Z) - Monte Carlo Game Solver [4.38602607138044]
It uses online learning of playout policies and Monte Carlo Tree Search.
The learned policy and the information in the Monte Carlo tree are used to order moves in game solvers.
arXiv Detail & Related papers (2020-01-15T00:20:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.