History Filtering in Imperfect Information Games: Algorithms and
Complexity
- URL: http://arxiv.org/abs/2311.14651v1
- Date: Fri, 24 Nov 2023 18:34:36 GMT
- Title: History Filtering in Imperfect Information Games: Algorithms and
Complexity
- Authors: Christopher Solinas, Douglas Rebstock, Nathan R. Sturtevant, Michael
Buro
- Abstract summary: We introduce and analyze the computational aspects and tractability of filtering histories for subgame decomposition.
We show that constructing a single history from the root of the subgame is generally intractable.
We also introduce a novel Markov Chain Monte Carlo-based generation algorithm for trick-taking card games.
- Score: 16.23892847804002
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Historically applied exclusively to perfect information games, depth-limited
search with value functions has been key to recent advances in AI for imperfect
information games. Most prominent approaches with strong theoretical guarantees
require subgame decomposition - a process in which a subgame is computed from
public information and player beliefs. However, subgame decomposition can
itself require non-trivial computations, and its tractability depends on the
existence of efficient algorithms for either full enumeration or generation of
the histories that form the root of the subgame. Despite this, no formal
analysis of the tractability of such computations has been established in prior
work, and application domains have often consisted of games, such as poker, for
which enumeration is trivial on modern hardware. Applying these ideas to more
complex domains requires understanding their cost.
In this work, we introduce and analyze the computational aspects and
tractability of filtering histories for subgame decomposition. We show that
constructing a single history from the root of the subgame is generally
intractable, and then provide a necessary and sufficient condition for
efficient enumeration. We also introduce a novel Markov Chain Monte Carlo-based
generation algorithm for trick-taking card games - a domain where enumeration
is often prohibitively expensive. Our experiments demonstrate its improved
scalability in the trick-taking card game Oh Hell. These contributions clarify
when and how depth-limited search via subgame decomposition can be an effective
tool for sequential decision-making in imperfect information settings.
Related papers
- Improve Value Estimation of Q Function and Reshape Reward with Monte Carlo Tree Search [0.4450107621124637]
Reinforcement learning has achieved remarkable success in perfect information games such as Go and Atari.
Research in reinforcement learning for imperfect information games has been relatively limited due to the more complex game structures and randomness.
In this paper, we focus on Uno, an imperfect information game, and aim to address these problems by reducing Q value overestimation and reshaping reward function.
arXiv Detail & Related papers (2024-10-15T14:31:54Z) - SPRING: Studying the Paper and Reasoning to Play Games [102.5587155284795]
We propose a novel approach, SPRING, to read the game's original academic paper and use the knowledge learned to reason and play the game through a large language model (LLM)
In experiments, we study the quality of in-context "reasoning" induced by different forms of prompts under the setting of the Crafter open-world environment.
Our experiments suggest that LLMs, when prompted with consistent chain-of-thought, have great potential in completing sophisticated high-level trajectories.
arXiv Detail & Related papers (2023-05-24T18:14:35Z) - The Update-Equivalence Framework for Decision-Time Planning [78.44953498421854]
We introduce an alternative framework for decision-time planning that is not based on solving subgames, but rather on update equivalence.
We derive a provably sound search algorithm for fully cooperative games based on mirror descent and a search algorithm for adversarial games based on magnetic mirror descent.
arXiv Detail & Related papers (2023-04-25T20:28:55Z) - Hardness of Independent Learning and Sparse Equilibrium Computation in
Markov Games [70.19141208203227]
We consider the problem of decentralized multi-agent reinforcement learning in Markov games.
We show that no algorithm attains no-regret in general-sum games when executed independently by all players.
We show that our lower bounds hold even for seemingly easier setting in which all agents are controlled by a centralized algorithm.
arXiv Detail & Related papers (2023-03-22T03:28:12Z) - Predicting Winning Regions in Parity Games via Graph Neural Networks
(Extended Abstract) [68.8204255655161]
We present an incomplete-time approach to determining the winning regions of parity games via graph neural networks.
It correctly determines the winning regions of $$60% of the games in our data set and only incurs minor errors in the remaining ones.
arXiv Detail & Related papers (2022-10-18T15:10:25Z) - Learning Correlated Equilibria in Mean-Field Games [62.14589406821103]
We develop the concepts of Mean-Field correlated and coarse-correlated equilibria.
We show that they can be efficiently learnt in emphall games, without requiring any additional assumption on the structure of the game.
arXiv Detail & Related papers (2022-08-22T08:31:46Z) - Public Information Representation for Adversarial Team Games [31.29335755664997]
adversarial team games reside in the asymmetric information available to the team members during the play.
Our algorithms convert a sequential team game with adversaries to a classical two-player zero-sum game.
Due to the NP-hard nature of the problem, the resulting Public Team game may be exponentially larger than the original one.
arXiv Detail & Related papers (2022-01-25T15:07:12Z) - Revisiting Game Representations: The Hidden Costs of Efficiency in
Sequential Decision-making Algorithms [0.6749750044497732]
Recent advancements in algorithms for sequential decision-making under imperfect information have shown remarkable success in large games.
These algorithms traditionally formalize the games using the extensive-form game formalism.
We show that a popular workaround involves using a specialized representation based on player specific information-state trees.
arXiv Detail & Related papers (2021-12-20T22:34:19Z) - From Poincar\'e Recurrence to Convergence in Imperfect Information
Games: Finding Equilibrium via Regularization [49.368421783733815]
We show how adapting the reward can give strong convergence guarantees in monotone games.
We also show how this reward adaptation technique can be leveraged to build algorithms that converge exactly to the Nash equilibrium.
arXiv Detail & Related papers (2020-02-19T21:36:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.