Learning Models of Individual Behavior in Chess
- URL: http://arxiv.org/abs/2008.10086v3
- Date: Wed, 15 Jun 2022 21:44:30 GMT
- Title: Learning Models of Individual Behavior in Chess
- Authors: Reid McIlroy-Young, Russell Wang, Siddhartha Sen, Jon Kleinberg,
Ashton Anderson
- Abstract summary: We develop highly accurate predictive models of individual human behavior in chess.
Our work demonstrates a way to bring AI systems into better alignment with the behavior of individual people.
- Score: 4.793072503820555
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: AI systems that can capture human-like behavior are becoming increasingly
useful in situations where humans may want to learn from these systems,
collaborate with them, or engage with them as partners for an extended
duration. In order to develop human-oriented AI systems, the problem of
predicting human actions -- as opposed to predicting optimal actions -- has
received considerable attention. Existing work has focused on capturing human
behavior in an aggregate sense, which potentially limits the benefit any
particular individual could gain from interaction with these systems. We extend
this line of work by developing highly accurate predictive models of individual
human behavior in chess. Chess is a rich domain for exploring human-AI
interaction because it combines a unique set of properties: AI systems achieved
superhuman performance many years ago, and yet humans still interact with them
closely, both as opponents and as preparation tools, and there is an enormous
corpus of recorded data on individual player games. Starting with Maia, an
open-source version of AlphaZero trained on a population of human players, we
demonstrate that we can significantly improve prediction accuracy of a
particular player's moves by applying a series of fine-tuning methods.
Furthermore, our personalized models can be used to perform stylometry --
predicting who made a given set of moves -- indicating that they capture human
decision-making at an individual level. Our work demonstrates a way to bring AI
systems into better alignment with the behavior of individual people, which
could lead to large improvements in human-AI interaction.
Related papers
- Maia-2: A Unified Model for Human-AI Alignment in Chess [10.577896749797485]
We propose a unified modeling approach for human-AI alignment in chess.
We introduce a skill-aware attention mechanism to dynamically integrate players' strengths with encoded chess positions.
Our results pave the way for deeper insights into human decision-making and AI-guided teaching tools.
arXiv Detail & Related papers (2024-09-30T17:54:23Z) - Explainable Human-AI Interaction: A Planning Perspective [32.477369282996385]
AI systems need to be explainable to the humans in the loop.
We will discuss how the AI agent can use mental models to either conform to human expectations, or change those expectations through explanatory communication.
While the main focus of the book is on cooperative scenarios, we will point out how the same mental models can be used for obfuscation and deception.
arXiv Detail & Related papers (2024-05-19T22:22:21Z) - SACSoN: Scalable Autonomous Control for Social Navigation [62.59274275261392]
We develop methods for training policies for socially unobtrusive navigation.
By minimizing this counterfactual perturbation, we can induce robots to behave in ways that do not alter the natural behavior of humans in the shared space.
We collect a large dataset where an indoor mobile robot interacts with human bystanders.
arXiv Detail & Related papers (2023-06-02T19:07:52Z) - Learning to Influence Human Behavior with Offline Reinforcement Learning [70.7884839812069]
We focus on influence in settings where there is a need to capture human suboptimality.
Experiments online with humans is potentially unsafe, and creating a high-fidelity simulator of the environment is often impractical.
We show that offline reinforcement learning can learn to effectively influence suboptimal humans by extending and combining elements of observed human-human behavior.
arXiv Detail & Related papers (2023-03-03T23:41:55Z) - HERD: Continuous Human-to-Robot Evolution for Learning from Human
Demonstration [57.045140028275036]
We show that manipulation skills can be transferred from a human to a robot through the use of micro-evolutionary reinforcement learning.
We propose an algorithm for multi-dimensional evolution path searching that allows joint optimization of both the robot evolution path and the policy.
arXiv Detail & Related papers (2022-12-08T15:56:13Z) - Detecting Individual Decision-Making Style: Exploring Behavioral
Stylometry in Chess [4.793072503820555]
We present a transformer-based approach to behavioral stylometry in the context of chess.
Our method operates in a few-shot classification framework, and can correctly identify a player from among thousands of candidate players.
We consider more broadly what our resulting embeddings reveal about human style in chess, as well as the potential ethical implications.
arXiv Detail & Related papers (2022-08-02T11:18:16Z) - Role of Human-AI Interaction in Selective Prediction [20.11364033416315]
We study the impact of communicating different types of information to humans about the AI system's decision to defer.
We show that it is possible to significantly boost human performance by informing the human of the decision to defer, but not revealing the prediction of the AI.
arXiv Detail & Related papers (2021-12-13T16:03:13Z) - Aligning Superhuman AI with Human Behavior: Chess as a Model System [5.236087378443016]
We develop Maia, a customized version of Alpha-Zero trained on human chess games, that predicts human moves at a much higher accuracy than existing engines.
For a dual task of predicting whether a human will make a large mistake on the next move, we develop a deep neural network that significantly outperforms competitive baselines.
arXiv Detail & Related papers (2020-06-02T18:12:52Z) - Learning to Complement Humans [67.38348247794949]
A rising vision for AI in the open world centers on the development of systems that can complement humans for perceptual, diagnostic, and reasoning tasks.
We demonstrate how an end-to-end learning strategy can be harnessed to optimize the combined performance of human-machine teams.
arXiv Detail & Related papers (2020-05-01T20:00:23Z) - Is the Most Accurate AI the Best Teammate? Optimizing AI for Teamwork [54.309495231017344]
We argue that AI systems should be trained in a human-centered manner, directly optimized for team performance.
We study this proposal for a specific type of human-AI teaming, where the human overseer chooses to either accept the AI recommendation or solve the task themselves.
Our experiments with linear and non-linear models on real-world, high-stakes datasets show that the most accuracy AI may not lead to highest team performance.
arXiv Detail & Related papers (2020-04-27T19:06:28Z) - Joint Inference of States, Robot Knowledge, and Human (False-)Beliefs [90.20235972293801]
Aiming to understand how human (false-temporal)-belief-a core socio-cognitive ability unify-would affect human interactions with robots, this paper proposes to adopt a graphical model to the representation of object states, robot knowledge, and human (false-)beliefs.
An inference algorithm is derived to fuse individual pg from all robots across multi-views into a joint pg, which affords more effective reasoning inference capability to overcome the errors originated from a single view.
arXiv Detail & Related papers (2020-04-25T23:02:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.