Neural Networks for Chess
- URL: http://arxiv.org/abs/2209.01506v1
- Date: Sat, 3 Sep 2022 22:17:16 GMT
- Title: Neural Networks for Chess
- Authors: Dominik Klein
- Abstract summary: AlphaZero, Leela Chess Zero and Stockfish NNUE revolutionized Computer Chess.
This book gives a complete introduction into the technical inner workings of such engines.
- Score: 2.055949720959582
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: AlphaZero, Leela Chess Zero and Stockfish NNUE revolutionized Computer Chess.
This book gives a complete introduction into the technical inner workings of
such engines. The book is split into four main chapters -- excluding chapter 1
(introduction) and chapter 6 (conclusion): Chapter 2 introduces neural networks
and covers all the basic building blocks that are used to build deep networks
such as those used by AlphaZero. Contents include the perceptron,
back-propagation and gradient descent, classification, regression, multilayer
perceptron, vectorization techniques, convolutional networks, squeeze and
excitation networks, fully connected networks, batch normalization and
rectified linear units, residual layers, overfitting and underfitting. Chapter
3 introduces classical search techniques used for chess engines as well as
those used by AlphaZero. Contents include minimax, alpha-beta search, and Monte
Carlo tree search. Chapter 4 shows how modern chess engines are designed. Aside
from the ground-breaking AlphaGo, AlphaGo Zero and AlphaZero we cover Leela
Chess Zero, Fat Fritz, Fat Fritz 2 and Efficiently Updatable Neural Networks
(NNUE) as well as Maia. Chapter 5 is about implementing a miniaturized
AlphaZero. Hexapawn, a minimalistic version of chess, is used as an example for
that. Hexapawn is solved by minimax search and training positions for
supervised learning are generated. Then as a comparison, an AlphaZero-like
training loop is implemented where training is done via self-play combined with
reinforcement learning. Finally, AlphaZero-like training and supervised
training are compared.
Related papers
- Enhancing Chess Reinforcement Learning with Graph Representation [21.919003715442074]
We introduce a more general architecture based on Graph Neural Networks (GNN)
We show that this new architecture outperforms previous architectures with a similar number of parameters.
We also show that the model, when trained on a smaller $5times 5$ variant of chess, is able to be quickly fine-tuned to play on regular $8times 8$ chess.
arXiv Detail & Related papers (2024-10-31T09:18:47Z) - MiniZero: Comparative Analysis of AlphaZero and MuZero on Go, Othello, and Atari Games [9.339645051415115]
MiniZero is a zero-knowledge learning framework that supports four state-of-the-art algorithms.
We evaluate the performance of each algorithm in two board games, 9x9 Go and 8x8 Othello, as well as 57 Atari games.
arXiv Detail & Related papers (2023-10-17T14:29:25Z) - NASiam: Efficient Representation Learning using Neural Architecture
Search for Siamese Networks [76.8112416450677]
Siamese networks are one of the most trending methods to achieve self-supervised visual representation learning (SSL)
NASiam is a novel approach that uses for the first time differentiable NAS to improve the multilayer perceptron projector and predictor (encoder/predictor pair)
NASiam reaches competitive performance in both small-scale (i.e., CIFAR-10/CIFAR-100) and large-scale (i.e., ImageNet) image classification datasets while costing only a few GPU hours.
arXiv Detail & Related papers (2023-01-31T19:48:37Z) - Evaluation Beyond Task Performance: Analyzing Concepts in AlphaZero in
Hex [39.001544338346655]
We investigate AlphaZero's internal representations in the game of Hex using two evaluation techniques from natural language processing (NLP): model probing and behavioral tests.
We find that concepts related to short-term end-game planning are best encoded in the final layers of the model, whereas concepts related to long-term planning are encoded in the middle layers of the model.
arXiv Detail & Related papers (2022-11-26T21:59:11Z) - An AlphaZero-Inspired Approach to Solving Search Problems [63.24965775030674]
We adapt the methods and techniques used in AlphaZero for solving search problems.
We describe possible representations in terms of easy-instance solvers and self-reductions.
We also describe a version of Monte Carlo tree search adapted for search problems.
arXiv Detail & Related papers (2022-07-02T23:39:45Z) - Training Your Sparse Neural Network Better with Any Mask [106.134361318518]
Pruning large neural networks to create high-quality, independently trainable sparse masks is desirable.
In this paper we demonstrate an alternative opportunity: one can customize the sparse training techniques to deviate from the default dense network training protocols.
Our new sparse training recipe is generally applicable to improving training from scratch with various sparse masks.
arXiv Detail & Related papers (2022-06-26T00:37:33Z) - AlphaZero-Inspired General Board Game Learning and Playing [0.0]
Recently, the seminal algorithms AlphaGo and AlphaZero have started a new era in game learning and deep reinforcement learning.
In this paper, we pick an important element of AlphaZero - the Monte Carlo Tree Search (MCTS) planning stage - and combine it with reinforcement learning (RL) agents.
We apply this new architecture to several complex games (Othello, ConnectFour, Rubik's Cube) and show the advantages achieved with this AlphaZero-inspired MCTS wrapper.
arXiv Detail & Related papers (2022-04-28T07:04:14Z) - Acquisition of Chess Knowledge in AlphaZero [14.41428465712717]
We show that human knowledge is acquired by the AlphaZero neural network as it trains on the game of chess.
By probing for a broad range of human chess concepts we show when and where these concepts are represented in the AlphaZero network.
We also provide a behavioural analysis focusing on opening play, including qualitative analysis from chess Grandmaster Vladimir Kramnik.
arXiv Detail & Related papers (2021-11-17T17:46:19Z) - Learning Compositional Neural Programs for Continuous Control [62.80551956557359]
We propose a novel solution to challenging sparse-reward, continuous control problems.
Our solution, dubbed AlphaNPI-X, involves three separate stages of learning.
We empirically show that AlphaNPI-X can effectively learn to tackle challenging sparse manipulation tasks.
arXiv Detail & Related papers (2020-07-27T08:27:14Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z) - Backward Feature Correction: How Deep Learning Performs Deep
(Hierarchical) Learning [66.05472746340142]
This paper analyzes how multi-layer neural networks can perform hierarchical learning _efficiently_ and _automatically_ by SGD on the training objective.
We establish a new principle called "backward feature correction", where the errors in the lower-level features can be automatically corrected when training together with the higher-level layers.
arXiv Detail & Related papers (2020-01-13T17:28:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.