Using Graph Convolutional Networks and TD($\lambda$) to play the game of
Risk
- URL: http://arxiv.org/abs/2009.06355v1
- Date: Thu, 10 Sep 2020 18:47:08 GMT
- Title: Using Graph Convolutional Networks and TD($\lambda$) to play the game of
Risk
- Authors: Jamie Carr
- Abstract summary: Risk is a 6 player game with significant randomness and a large game-tree complexity.
Previous AIs focus on creating high-level handcrafted features determine agent decision making.
I create D.A.D, A Risk agent using temporal difference reinforcement learning to train a Deep Neural Network.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Risk is 6 player game with significant randomness and a large game-tree
complexity which poses a challenge to creating an agent to play the game
effectively. Previous AIs focus on creating high-level handcrafted features
determine agent decision making. In this project, I create D.A.D, A Risk agent
using temporal difference reinforcement learning to train a Deep Neural Network
including a Graph Convolutional Network to evaluate player positions. This is
used in a game-tree to select optimal moves. This allows minimal handcrafting
of knowledge into the AI, assuring input features are as low-level as possible
to allow the network to extract useful and sophisticated features itself, even
with the network starting from a random initialisation. I also tackle the issue
of non-determinism in Risk by introducing a new method of interpreting attack
moves necessary for the search. The result is an AI which wins 35% of the time
versus 5 of best inbuilt AIs in Lux Delux, a Risk variant.
Related papers
- Toward Human-AI Alignment in Large-Scale Multi-Player Games [24.784173202415687]
We analyze extensive human gameplay data from Xbox's Bleeding Edge (100K+ games)
We find that while human players exhibit variability in fight-flight and explore-exploit behavior, AI players tend towards uniformity.
These stark differences underscore the need for interpretable evaluation, design, and integration of AI in human-aligned applications.
arXiv Detail & Related papers (2024-02-05T22:55:33Z) - DanZero+: Dominating the GuanDan Game through Reinforcement Learning [95.90682269990705]
We develop an AI program for an exceptionally complex and popular card game called GuanDan.
We first put forward an AI program named DanZero for this game.
In order to further enhance the AI's capabilities, we apply policy-based reinforcement learning algorithm to GuanDan.
arXiv Detail & Related papers (2023-12-05T08:07:32Z) - Explaining How a Neural Network Play the Go Game and Let People Learn [26.192580802652742]
The AI model has surpassed human players in the game of Go.
It is widely believed that the AI model has encoded new knowledge about the Go game beyond human players.
arXiv Detail & Related papers (2023-10-15T13:57:50Z) - Are AlphaZero-like Agents Robust to Adversarial Perturbations? [73.13944217915089]
AlphaZero (AZ) has demonstrated that neural-network-based Go AIs can surpass human performance by a large margin.
We ask whether adversarial states exist for Go AIs that may lead them to play surprisingly wrong actions.
We develop the first adversarial attack on Go AIs that can efficiently search for adversarial states by strategically reducing the search space.
arXiv Detail & Related papers (2022-11-07T18:43:25Z) - Mastering the Game of Stratego with Model-Free Multiagent Reinforcement
Learning [86.37438204416435]
Stratego is one of the few iconic board games that Artificial Intelligence (AI) has not yet mastered.
Decisions in Stratego are made over a large number of discrete actions with no obvious link between action and outcome.
DeepNash beats existing state-of-the-art AI methods in Stratego and achieved a yearly (2022) and all-time top-3 rank on the Gravon games platform.
arXiv Detail & Related papers (2022-06-30T15:53:19Z) - The Feasibility and Inevitability of Stealth Attacks [63.14766152741211]
We study new adversarial perturbations that enable an attacker to gain control over decisions in generic Artificial Intelligence systems.
In contrast to adversarial data modification, the attack mechanism we consider here involves alterations to the AI system itself.
arXiv Detail & Related papers (2021-06-26T10:50:07Z) - Model-Free Online Learning in Unknown Sequential Decision Making
Problems and Games [114.90723492840499]
In large two-player zero-sum imperfect-information games, modern extensions of counterfactual regret minimization (CFR) are currently the practical state of the art for computing a Nash equilibrium.
We formalize an online learning setting in which the strategy space is not known to the agent.
We give an efficient algorithm that achieves $O(T3/4)$ regret with high probability for that setting, even when the agent faces an adversarial environment.
arXiv Detail & Related papers (2021-03-08T04:03:24Z) - AI solutions for drafting in Magic: the Gathering [0.0]
We present a dataset of over 100,000 simulated, anonymized human drafts collected from Draftsim.com.
We propose four diverse strategies for drafting agents, including a primitive drafting agent, an expert-tuned complex agent, a Naive Bayes agent, and a deep neural network agent.
This work helps to identify next steps in the creation of humanlike drafting agents, and can serve as a benchmark for the next generation of drafting bots.
arXiv Detail & Related papers (2020-09-01T18:44:10Z) - Playing Catan with Cross-dimensional Neural Network [0.0]
It is challenging to build AI agents by Reinforcement Learning (RL for short) without domain knowledge nors.
In this paper, we introduce cross-dimensional neural networks to handle a mixture of information sources and a wide variety of outputs, and empirically demonstrate that the network dramatically improves RL in Catan.
We also show that, for the first time, a RL agent can outperform jsettler, the best agent available.
arXiv Detail & Related papers (2020-08-17T04:09:29Z) - Learning to Play Sequential Games versus Unknown Opponents [93.8672371143881]
We consider a repeated sequential game between a learner, who plays first, and an opponent who responds to the chosen action.
We propose a novel algorithm for the learner when playing against an adversarial sequence of opponents.
Our results include algorithm's regret guarantees that depend on the regularity of the opponent's response.
arXiv Detail & Related papers (2020-07-10T09:33:05Z) - Testing match-3 video games with Deep Reinforcement Learning [0.0]
We study the possibility to use the Deep Reinforcement Learning to automate the testing process in match-3 video games.
We test this kind of network on the Jelly Juice game, a match-3 video game developed by the redBit Games.
arXiv Detail & Related papers (2020-06-30T12:41:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.