Monte Carlo Game Solver
- URL: http://arxiv.org/abs/2001.05087v1
- Date: Wed, 15 Jan 2020 00:20:13 GMT
- Title: Monte Carlo Game Solver
- Authors: Tristan Cazenave
- Abstract summary: It uses online learning of playout policies and Monte Carlo Tree Search.
The learned policy and the information in the Monte Carlo tree are used to order moves in game solvers.
- Score: 4.38602607138044
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a general algorithm to order moves so as to speedup exact game
solvers. It uses online learning of playout policies and Monte Carlo Tree
Search. The learned policy and the information in the Monte Carlo tree are used
to order moves in game solvers. They improve greatly the solving time for
multiple games.
Related papers
- Deep Reinforcement Learning for 5*5 Multiplayer Go [6.222520876209623]
We propose to use and analyze the latest algorithms that use search and Deep Reinforcement Learning (DRL)
We show that using search and DRL we were able to improve the level of play, even though there are more than two players.
arXiv Detail & Related papers (2024-05-23T07:44:24Z) - Monte Carlo Search Algorithms Discovering Monte Carlo Tree Search Exploration Terms [4.561007128508218]
The optimized Monte Carlo Tree Search algorithms are PUCT and SHUSS.
For small search budgets of 32 evaluations the discovered root exploration terms make both algorithms competitive.
arXiv Detail & Related papers (2024-04-14T17:06:20Z) - Monte-Carlo Tree Search for Multi-Agent Pathfinding: Preliminary Results [60.4817465598352]
We introduce an original variant of Monte-Carlo Tree Search (MCTS) tailored to multi-agent pathfinding.
Specifically, we use individual paths to assist the agents with the the goal-reaching behavior.
We also use a dedicated decomposition technique to reduce the branching factor of the tree search procedure.
arXiv Detail & Related papers (2023-07-25T12:33:53Z) - Solving Royal Game of Ur Using Reinforcement Learning [0.0]
We train our agents using different methods namely Monte Carlo, Qlearning and Expected Sarsa to learn optimal policy to play the strategic Royal Game of Ur.
Although it is hard to conclude that when trained with limited resources which algorithm performs better overall, but Expected Sarsa shows promising results when it comes to fastest learning.
arXiv Detail & Related papers (2022-08-23T01:26:37Z) - An AlphaZero-Inspired Approach to Solving Search Problems [63.24965775030674]
We adapt the methods and techniques used in AlphaZero for solving search problems.
We describe possible representations in terms of easy-instance solvers and self-reductions.
We also describe a version of Monte Carlo tree search adapted for search problems.
arXiv Detail & Related papers (2022-07-02T23:39:45Z) - Learning in Mean Field Games: A Survey [44.93300994923148]
Mean Field Games (MFGs) rely on a mean-field approximation to allow the number of players to grow to infinity.
Recent literature on Reinforcement Learning methods to learnlibria and social optima in MFGs.
We present a general framework for classical iterative methods to solve MFGs in an exact way.
arXiv Detail & Related papers (2022-05-25T17:49:37Z) - MCTS Based Agents for Multistage Single-Player Card Game [0.0]
The article presents the use of Monte Carlo Tree Search algorithms for the card game Lord of the Rings.
The main challenge was the complexity of the game mechanics, in which each round consists of 5 decision stages and 2 random stages.
To test various decision-making algorithms, a game simulator has been implemented.
arXiv Detail & Related papers (2021-09-24T10:56:54Z) - Generating Diverse and Competitive Play-Styles for Strategy Games [58.896302717975445]
We propose Portfolio Monte Carlo Tree Search with Progressive Unpruning for playing a turn-based strategy game (Tribes)
We show how it can be parameterized so a quality-diversity algorithm (MAP-Elites) is used to achieve different play-styles while keeping a competitive level of play.
Our results show that this algorithm is capable of achieving these goals even for an extensive collection of game levels beyond those used for training.
arXiv Detail & Related papers (2021-04-17T20:33:24Z) - Monte-Carlo Tree Search as Regularized Policy Optimization [47.541849128047865]
We show that AlphaZero's search algorithms are an approximation to the solution of a specific regularized policy optimization problem.
We propose a variant of AlphaZero which uses the exact solution to this policy optimization problem, and show experimentally that it reliably outperforms the original algorithm in multiple domains.
arXiv Detail & Related papers (2020-07-24T13:01:34Z) - Competing in a Complex Hidden Role Game with Information Set Monte Carlo
Tree Search [0.0]
The Information Set Monte Carlo Tree Search (ISMCTS) family of algorithms outperforms previous algorithms using Monte Carlo methods in imperfect information games.
This paper is applied to Secret Hitler, a popular social deduction board game that combines traditional hidden role mechanics with the randomness of a card deck.
arXiv Detail & Related papers (2020-05-14T17:21:10Z) - From Poincar\'e Recurrence to Convergence in Imperfect Information
Games: Finding Equilibrium via Regularization [49.368421783733815]
We show how adapting the reward can give strong convergence guarantees in monotone games.
We also show how this reward adaptation technique can be leveraged to build algorithms that converge exactly to the Nash equilibrium.
arXiv Detail & Related papers (2020-02-19T21:36:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.