Finite Group Equivariant Neural Networks for Games
- URL: http://arxiv.org/abs/2009.05027v1
- Date: Thu, 10 Sep 2020 17:46:09 GMT
- Title: Finite Group Equivariant Neural Networks for Games
- Authors: Ois\'in Carroll, Joeran Beel
- Abstract summary: Group equivariant CNNs in existing work create networks which can exploit symmetries to improve learning.
We introduce Finite Group Neural Networks (FGNNs), a method for creating agents with an innate understanding of these board positions.
FGNNs are shown to improve the performance of networks playing checkers (draughts) and can be easily adapted to other games and learning problems.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Games such as go, chess and checkers have multiple equivalent game states,
i.e. multiple board positions where symmetrical and opposite moves should be
made. These equivalences are not exploited by current state of the art neural
agents which instead must relearn similar information, thereby wasting
computing time. Group equivariant CNNs in existing work create networks which
can exploit symmetries to improve learning, however, they lack the
expressiveness to correctly reflect the move embeddings necessary for games. We
introduce Finite Group Neural Networks (FGNNs), a method for creating agents
with an innate understanding of these board positions. FGNNs are shown to
improve the performance of networks playing checkers (draughts), and can be
easily adapted to other games and learning problems. Additionally, FGNNs can be
created from existing network architectures. These include, for the first time,
those with skip connections and arbitrary layer types. We demonstrate that an
equivariant version of U-Net (FGNN-U-Net) outperforms the unmodified network in
image segmentation.
Related papers
- Enhancing lattice kinetic schemes for fluid dynamics with Lattice-Equivariant Neural Networks [79.16635054977068]
We present a new class of equivariant neural networks, dubbed Lattice-Equivariant Neural Networks (LENNs)
Our approach develops within a recently introduced framework aimed at learning neural network-based surrogate models Lattice Boltzmann collision operators.
Our work opens towards practical utilization of machine learning-augmented Lattice Boltzmann CFD in real-world simulations.
arXiv Detail & Related papers (2024-05-22T17:23:15Z) - Degree-based stratification of nodes in Graph Neural Networks [66.17149106033126]
We modify the Graph Neural Network (GNN) architecture so that the weight matrices are learned, separately, for the nodes in each group.
This simple-to-implement modification seems to improve performance across datasets and GNN methods.
arXiv Detail & Related papers (2023-12-16T14:09:23Z) - From Images to Connections: Can DQN with GNNs learn the Strategic Game
of Hex? [22.22813915303447]
We investigate whether graph neural networks (GNNs) can replace convolutional neural networks (CNNs) in self-play reinforcement learning.
GNNs excel at dealing with long range dependency situations in game states and are less prone to overfitting.
This suggests a potential paradigm shift, signaling the use of game-specific structures to reshape self-play reinforcement learning.
arXiv Detail & Related papers (2023-11-22T14:20:15Z) - Using Logic Programming and Kernel-Grouping for Improving
Interpretability of Convolutional Neural Networks [1.6317061277457001]
We present a neurosymbolic framework, NeSyFOLD-G that generates a symbolic rule-set using the last layer kernels of the CNN.
We show that grouping similar kernels leads to a significant reduction in the size of the rule-set generated by FOLD-SE-M.
We also propose a novel algorithm for labeling each predicate in the rule-set with the semantic concept(s) that its corresponding kernel group represents.
arXiv Detail & Related papers (2023-10-19T18:12:49Z) - You Can Have Better Graph Neural Networks by Not Training Weights at
All: Finding Untrained GNNs Tickets [105.24703398193843]
Untrainedworks in graph neural networks (GNNs) still remains mysterious.
We show that the found untrainedworks can substantially mitigate the GNN over-smoothing problem.
We also observe that such sparse untrainedworks have appealing performance in out-of-distribution detection and robustness of input perturbations.
arXiv Detail & Related papers (2022-11-28T14:17:36Z) - A General Framework For Proving The Equivariant Strong Lottery Ticket
Hypothesis [15.376680573592997]
Modern neural networks are capable of incorporating more than just translation symmetry.
We generalize the Strong Lottery Ticket Hypothesis (SLTH) to functions that preserve the action of the group $G$.
We prove our theory by overparametrized $textE(2)$-steerable CNNs and message passing GNNs.
arXiv Detail & Related papers (2022-06-09T04:40:18Z) - mGNN: Generalizing the Graph Neural Networks to the Multilayer Case [0.0]
We propose mGNN, a framework meant to generalize GNNs to multi-layer networks.
Our approach is general (i.e., not task specific) and has the advantage of extending any type of GNN without any computational overhead.
We test the framework into three different tasks (node and network classification, link prediction) to validate it.
arXiv Detail & Related papers (2021-09-21T12:02:12Z) - Towards Deeper Graph Neural Networks with Differentiable Group
Normalization [61.20639338417576]
Graph neural networks (GNNs) learn the representation of a node by aggregating its neighbors.
Over-smoothing is one of the key issues which limit the performance of GNNs as the number of layers increases.
We introduce two over-smoothing metrics and a novel technique, i.e., differentiable group normalization (DGN)
arXiv Detail & Related papers (2020-06-12T07:18:02Z) - Wide and Deep Graph Neural Networks with Distributed Online Learning [175.96910854433574]
Graph neural networks (GNNs) learn representations from network data with naturally distributed architectures.
Online learning can be used to retrain GNNs at testing time, overcoming this issue.
This paper proposes the Wide and Deep GNN (WD-GNN), a novel architecture that can be easily updated with distributed online learning mechanisms.
arXiv Detail & Related papers (2020-06-11T12:48:03Z) - EdgeNets:Edge Varying Graph Neural Networks [179.99395949679547]
This paper puts forth a general framework that unifies state-of-the-art graph neural networks (GNNs) through the concept of EdgeNet.
An EdgeNet is a GNN architecture that allows different nodes to use different parameters to weigh the information of different neighbors.
This is a general linear and local operation that a node can perform and encompasses under one formulation all existing graph convolutional neural networks (GCNNs) as well as graph attention networks (GATs)
arXiv Detail & Related papers (2020-01-21T15:51:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.