Generalization over different cellular automata rules learned by a deep
feed-forward neural network
- URL: http://arxiv.org/abs/2103.14886v1
- Date: Sat, 27 Mar 2021 12:12:07 GMT
- Title: Generalization over different cellular automata rules learned by a deep
feed-forward neural network
- Authors: Marcel Aach, Jens Henrik Goebbert, Jenia Jitsev
- Abstract summary: A deep convolutional encoder-decoder network with short and long range skip connections is trained on various generated trajectories to predict the next CA state.
Results show that the network is able to learn the rules of various, complex cellular automata and generalize to unseen configurations.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To test generalization ability of a class of deep neural networks, we
randomly generate a large number of different rule sets for 2-D cellular
automata (CA), based on John Conway's Game of Life. Using these rules, we
compute several trajectories for each CA instance. A deep convolutional
encoder-decoder network with short and long range skip connections is trained
on various generated CA trajectories to predict the next CA state given its
previous states. Results show that the network is able to learn the rules of
various, complex cellular automata and generalize to unseen configurations. To
some extent, the network shows generalization to rule sets and neighborhood
sizes that were not seen during the training at all.
Related papers
- Generalization emerges from local optimization in a self-organized learning network [0.0]
We design and analyze a new paradigm for building supervised learning networks, driven only by local optimization rules without relying on a global error function.
Our network stores new knowledge in the nodes accurately and instantaneously, in the form of a lookup table.
We show on numerous examples of classification tasks that the networks generated by our algorithm systematically reach such a state of perfect generalization when the number of learned examples becomes sufficiently large.
We report on the dynamics of the change of state and show that it is abrupt and has the distinctive characteristics of a first order phase transition, a phenomenon already observed for traditional learning networks and known as grokking.
arXiv Detail & Related papers (2024-10-03T15:32:08Z) - Convolutional Neural Networks for Automated Cellular Automaton Classification [0.0]
We implement computer vision techniques to perform an automated classification of elementary cellular automata into the five Li-Packard classes.
We first show that previously developed deep learning approaches have in fact been trained to identify the local update rule.
We then present a convolutional neural network that performs nearly perfectly at identifying the behavioural class.
arXiv Detail & Related papers (2024-09-04T14:21:00Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Pathfinding Neural Cellular Automata [23.831530224401575]
Pathfinding is an important sub-component of a broad range of complex AI tasks, such as robot path planning, transport routing, and game playing.
We hand-code and learn models for Breadth-First Search (BFS), i.e. shortest path finding.
We present a neural implementation of Depth-First Search (DFS), and outline how it can be combined with neural BFS to produce an NCA for computing diameter of a graph.
We experiment with architectural modifications inspired by these hand-coded NCAs, training networks from scratch to solve the diameter problem on grid mazes while exhibiting strong ability generalization
arXiv Detail & Related papers (2023-01-17T11:45:51Z) - On the Effective Number of Linear Regions in Shallow Univariate ReLU
Networks: Convergence Guarantees and Implicit Bias [50.84569563188485]
We show that gradient flow converges in direction when labels are determined by the sign of a target network with $r$ neurons.
Our result may already hold for mild over- parameterization, where the width is $tildemathcalO(r)$ and independent of the sample size.
arXiv Detail & Related papers (2022-05-18T16:57:10Z) - Learning Graph Cellular Automata [25.520299226767946]
We focus on a generalised version of typical cellular automata (GCA)
In particular, we extend previous work that used convolutional neural networks to learn the transition rule of conventional GCA.
We show that it can represent any arbitrary GCA with finite and discrete state space.
arXiv Detail & Related papers (2021-10-27T07:42:48Z) - The Separation Capacity of Random Neural Networks [78.25060223808936]
We show that a sufficiently large two-layer ReLU-network with standard Gaussian weights and uniformly distributed biases can solve this problem with high probability.
We quantify the relevant structure of the data in terms of a novel notion of mutual complexity.
arXiv Detail & Related papers (2021-07-31T10:25:26Z) - Lattice gauge equivariant convolutional neural networks [0.0]
We propose Lattice gauge equivariant Convolutional Neural Networks (L-CNNs) for generic machine learning applications.
We show that L-CNNs can learn and generalize gauge invariant quantities that traditional convolutional neural networks are incapable of finding.
arXiv Detail & Related papers (2020-12-23T19:00:01Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z) - Learn to Predict Sets Using Feed-Forward Neural Networks [63.91494644881925]
This paper addresses the task of set prediction using deep feed-forward neural networks.
We present a novel approach for learning to predict sets with unknown permutation and cardinality.
We demonstrate the validity of our set formulations on relevant vision problems.
arXiv Detail & Related papers (2020-01-30T01:52:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.