Knowledge-Based Paranoia Search in Trick-Taking
- URL: http://arxiv.org/abs/2104.05423v1
- Date: Wed, 7 Apr 2021 09:12:45 GMT
- Title: Knowledge-Based Paranoia Search in Trick-Taking
- Authors: Stefan Edelkamp
- Abstract summary: This paper proposes emphknowledge-based paraonoia search (KBPS) to find forced wins during trick-taking in the card game Skat.
It combines efficient partial information game-tree search with knowledge representation and reasoning.
- Score: 1.3706331473063877
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes \emph{knowledge-based paraonoia search} (KBPS) to find
forced wins during trick-taking in the card game Skat; for some one of the most
interesting card games for three players. It combines efficient partial
information game-tree search with knowledge representation and reasoning. This
worst-case analysis, initiated after a small number of tricks, leads to a
prioritized choice of cards. We provide variants of KBPS for the declarer and
the opponents, and an approximation to find a forced win against most worlds in
the belief space. Replaying thousands of expert games, our evaluation indicates
that the AIs with the new algorithms perform better than humans in their play,
achieving an average score of over 1,000 points in the agreed standard for
evaluating Skat tournaments, the extended Seeger system.
Related papers
- People use fast, goal-directed simulation to reason about novel games [75.25089384921557]
We study how people reason about a range of simple but novel connect-n style board games.
We ask people to judge how fair and how fun the games are from very little experience.
arXiv Detail & Related papers (2024-07-19T07:59:04Z) - AlphaZero Gomoku [9.434566356382529]
We broaden the use of AlphaZero to Gomoku, an age-old tactical board game also referred to as "Five in a Row"
Our tests demonstrate AlphaZero's versatility in adapting to games other than Go.
arXiv Detail & Related papers (2023-09-04T00:20:06Z) - Targeted Search Control in AlphaZero for Effective Policy Improvement [93.30151539224144]
We introduce Go-Exploit, a novel search control strategy for AlphaZero.
Go-Exploit samples the start state of its self-play trajectories from an archive of states of interest.
Go-Exploit learns with a greater sample efficiency than standard AlphaZero.
arXiv Detail & Related papers (2023-02-23T22:50:24Z) - Optimisation of MCTS Player for The Lord of the Rings: The Card Game [0.0]
The article presents research on the use of Monte-Carlo Tree Search (MCTS) methods to create an artificial player for the popular card game "The Lord of the Rings"
arXiv Detail & Related papers (2021-09-24T14:42:32Z) - Generating Diverse and Competitive Play-Styles for Strategy Games [58.896302717975445]
We propose Portfolio Monte Carlo Tree Search with Progressive Unpruning for playing a turn-based strategy game (Tribes)
We show how it can be parameterized so a quality-diversity algorithm (MAP-Elites) is used to achieve different play-styles while keeping a competitive level of play.
Our results show that this algorithm is capable of achieving these goals even for an extensive collection of game levels beyond those used for training.
arXiv Detail & Related papers (2021-04-17T20:33:24Z) - On the Power of Refined Skat Selection [1.3706331473063877]
Skat is a fascinating card game, show-casing many of the intrinsic challenges for modern AI systems.
We propose hard expert-rules and a scoring function based on refined skat evaluation features.
Experiments emphasize the impact of the refined skat putting algorithm on the playing performance of the bots.
arXiv Detail & Related papers (2021-04-07T08:54:58Z) - Collaborative Agent Gameplay in the Pandemic Board Game [3.223284371460913]
Pandemic is an exemplar collaborative board game where all players coordinate to overcome challenges posed by events occurring during the game's progression.
This paper proposes an artificial agent which controls all players' actions and balances chances of winning versus risk of losing in this highly Evolutionary environment.
Results show that the proposed algorithm can find winning strategies more consistently in different games of varying difficulty.
arXiv Detail & Related papers (2021-03-21T13:18:20Z) - Faster Algorithms for Optimal Ex-Ante Coordinated Collusive Strategies
in Extensive-Form Zero-Sum Games [123.76716667704625]
We focus on the problem of finding an optimal strategy for a team of two players that faces an opponent in an imperfect-information zero-sum extensive-form game.
In that setting, it is known that the best the team can do is sample a profile of potentially randomized strategies (one per player) from a joint (a.k.a. correlated) probability distribution at the beginning of the game.
We provide an algorithm that computes such an optimal distribution by only using profiles where only one of the team members gets to randomize in each profile.
arXiv Detail & Related papers (2020-09-21T17:51:57Z) - Multi-Armed Bandits for Minesweeper: Profiting from
Exploration-Exploitation Synergy [0.0]
A popular computer puzzle, the game of Minesweeper requires its human players to have a mix of both luck and strategy to succeed.
We develop a novel methodology based on Reinforcement Learning to tackle the problem presented by this game.
arXiv Detail & Related papers (2020-07-25T01:44:50Z) - Learning to Play Sequential Games versus Unknown Opponents [93.8672371143881]
We consider a repeated sequential game between a learner, who plays first, and an opponent who responds to the chosen action.
We propose a novel algorithm for the learner when playing against an adversarial sequence of opponents.
Our results include algorithm's regret guarantees that depend on the regularity of the opponent's response.
arXiv Detail & Related papers (2020-07-10T09:33:05Z) - Efficient exploration of zero-sum stochastic games [83.28949556413717]
We investigate the increasingly important and common game-solving setting where we do not have an explicit description of the game but only oracle access to it through gameplay.
During a limited-duration learning phase, the algorithm can control the actions of both players in order to try to learn the game and how to play it well.
Our motivation is to quickly learn strategies that have low exploitability in situations where evaluating the payoffs of a queried strategy profile is costly.
arXiv Detail & Related papers (2020-02-24T20:30:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.