The Value of Chess Squares
- URL: http://arxiv.org/abs/2307.05330v2
- Date: Tue, 10 Oct 2023 00:35:45 GMT
- Title: The Value of Chess Squares
- Authors: Aditya Gupta and Shiva Maharaj and Nicholas Polson and Vadim Sokolov
- Abstract summary: Our model takes a triplet (Color, Piece, Square) as an input and calculates a value that measures the advantage/disadvantage of having this piece on this square.
Our methods build on recent advances in chess AI, and can accurately assess the worth of positions in a game of chess.
- Score: 5.647533385886476
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: We propose a neural network-based approach to calculate the value of a chess
square-piece combination. Our model takes a triplet (Color, Piece, Square) as
an input and calculates a value that measures the advantage/disadvantage of
having this piece on this square. Our methods build on recent advances in chess
AI, and can accurately assess the worth of positions in a game of chess. The
conventional approach assigns fixed values to pieces $(\symking=\infty,
\symqueen=9, \symrook=5, \symbishop=3, \symknight=3, \sympawn=1)$. We enhance
this analysis by introducing marginal valuations. We use deep Q-learning to
estimate the parameters of our model. We demonstrate our method by examining
the positioning of Knights and Bishops, and also provide valuable insights into
the valuation of pawns. Finally, we conclude by suggesting potential avenues
for future research.
Related papers
- Nearly Optimal Algorithms for Contextual Dueling Bandits from Adversarial Feedback [58.66941279460248]
Learning from human feedback plays an important role in aligning generative models, such as large language models (LLM)
We study a model within this problem domain--contextual dueling bandits with adversarial feedback, where the true preference label can be flipped by an adversary.
We propose an algorithm namely robust contextual dueling bandit (algo), which is based on uncertainty-weighted maximum likelihood estimation.
arXiv Detail & Related papers (2024-04-16T17:59:55Z) - Amortized Planning with Large-Scale Transformers: A Case Study on Chess [11.227110138932442]
This paper uses chess, a landmark planning problem in AI, to assess performance on a planning task.
ChessBench is a large-scale benchmark of 10 million chess games with legal move and value annotations (15 billion points) provided by Stockfish.
We show that, although a remarkably good approximation can be distilled into large-scale transformers via supervised learning, perfect distillation is still beyond reach.
arXiv Detail & Related papers (2024-02-07T00:36:24Z) - End-to-End Chess Recognition [11.15543089335477]
Current approaches use a pipeline of separate, independent, modules such as chessboard detection, square localization, and piece classification.
We explore an end-to-end approach to directly predict the configuration from the image, thus avoiding the error accumulation of the sequential approaches.
In contrast to existing datasets that are synthetically rendered and have only limited angles, ChessReD has photographs captured from various angles using smartphone cameras.
Our approach in chess recognition on the introduced challenging benchmark dataset outperforms related approaches, successfully recognizing the chess pieces' configuration in 15.26% of ChessReD's test images.
arXiv Detail & Related papers (2023-10-06T08:30:20Z) - Statistical analysis of chess games: space control and tipping points [0.0]
We first focus on spatial properties and the location of pieces and show that the number of possible moves during a game is positively correlated with its outcome.
We then study heatmaps of pieces and show that the spatial distribution of pieces varies less between human players than with engines (such as Stockfish)
arXiv Detail & Related papers (2023-04-22T14:54:34Z) - Representation Learning for General-sum Low-rank Markov Games [63.119870889883224]
We study multi-agent general-sum Markov games with nonlinear function approximation.
We focus on low-rank Markov games whose transition matrix admits a hidden low-rank structure on top of an unknown non-linear representation.
arXiv Detail & Related papers (2022-10-30T22:58:22Z) - There is no Accuracy-Interpretability Tradeoff in Reinforcement Learning
for Mazes [64.05903267230467]
Interpretability is an essential building block for trustworthiness in reinforcement learning systems.
We show that in certain cases, one can achieve policy interpretability while maintaining its optimality.
arXiv Detail & Related papers (2022-06-09T04:23:26Z) - Towards General Function Approximation in Zero-Sum Markov Games [126.58493169301012]
This paper considers two-player zero-sum finite-horizon Markov games with simultaneous moves.
Provably efficient algorithms for both decoupled and coordinated settings are developed.
arXiv Detail & Related papers (2021-07-30T15:25:13Z) - Determining Chess Game State From an Image [19.06796946564999]
This paper puts forth a new dataset synthesised from a 3D model that is an order of magnitude larger than existing ones.
A novel end-to-end chess recognition system is presented that combines traditional computer vision techniques with deep learning.
The described system achieves an error rate of 0.23% per square on the test set, 28 times better than the current state of the art.
arXiv Detail & Related papers (2021-04-30T13:02:13Z) - Almost Optimal Algorithms for Two-player Markov Games with Linear
Function Approximation [92.99933928528797]
We study reinforcement learning for two-player zero-sum Markov games with simultaneous moves.
We propose an algorithm Nash-UCRL-VTR based on the principle "Optimism-in-Face-of-Uncertainty"
We show that Nash-UCRL-VTR can provably achieve an $tildeO(dHsqrtT)$ regret, where $d$ is the linear function dimension.
arXiv Detail & Related papers (2021-02-15T09:09:16Z) - Playing Chess with Limited Look Ahead [0.0]
We train a deep neural network to serve as a static evaluation function.
We show that our static evaluation function has encoded some semblance of look ahead knowledge.
We show that, despite strict restrictions on look ahead depth, our engine recommends moves of equal strength in roughly $83%$ of our sample positions.
arXiv Detail & Related papers (2020-07-04T16:02:43Z) - Learning Zero-Sum Simultaneous-Move Markov Games Using Function
Approximation and Correlated Equilibrium [116.56359444619441]
We develop provably efficient reinforcement learning algorithms for two-player zero-sum finite-horizon Markov games.
In the offline setting, we control both players and aim to find the Nash Equilibrium by minimizing the duality gap.
In the online setting, we control a single player playing against an arbitrary opponent and aim to minimize the regret.
arXiv Detail & Related papers (2020-02-17T17:04:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.