On The Expressivity of Recurrent Neural Cascades
- URL: http://arxiv.org/abs/2312.09048v1
- Date: Thu, 14 Dec 2023 15:47:26 GMT
- Title: On The Expressivity of Recurrent Neural Cascades
- Authors: Nadezda Alexandrovna Knorozova, Alessandro Ronca
- Abstract summary: Recurrent Neural Cascades (RNCs) are the recurrent neural networks with no cyclic dependencies among recurrent neurons.
We show that RNCs can achieve the expressivity of all regular languages by introducing neurons that can implement groups.
- Score: 53.397276621815614
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Recurrent Neural Cascades (RNCs) are the recurrent neural networks with no
cyclic dependencies among recurrent neurons. This class of recurrent networks
has received a lot of attention in practice. Besides training methods for a
fixed architecture such as backpropagation, the cascade architecture naturally
allows for constructive learning methods, where recurrent nodes are added
incrementally one at a time, often yielding smaller networks. Furthermore,
acyclicity amounts to a structural prior that even for the same number of
neurons yields a more favourable sample complexity compared to a
fully-connected architecture. A central question is whether the advantages of
the cascade architecture come at the cost of a reduced expressivity. We provide
new insights into this question. We show that the regular languages captured by
RNCs with sign and tanh activation with positive recurrent weights are the
star-free regular languages. In order to establish our results we developed a
novel framework where capabilities of RNCs are accessed by analysing which
semigroups and groups a single neuron is able to implement. A notable
implication of our framework is that RNCs can achieve the expressivity of all
regular languages by introducing neurons that can implement groups.
Related papers
- On the Expressivity of Recurrent Neural Cascades with Identity [48.87943990557107]
We establish a close structural correspondence between RNC+ and semiautomata cascades.
A notable consequence of this result is that RNC+ are no more succinct than cascades of three-state semiautomata.
arXiv Detail & Related papers (2024-05-19T20:06:38Z) - De-novo Chemical Reaction Generation by Means of Temporal Convolutional
Neural Networks [3.357271554042638]
We present here a combination of two networks, Recurrent Neural Networks (RNN) and Temporarily Convolutional Neural Networks (TCN)
Recurrent Neural Networks are known for their autoregressive properties and are frequently used in language modelling with direct application to SMILES generation.
The relatively novel TCNs possess similar properties with wide receptive field while obeying the causality required for natural language processing (NLP)
It is shown that different fine-tuning protocols have a profound impact on generative scope of the model when applied on a dataset of interest via transfer learning.
arXiv Detail & Related papers (2023-10-26T12:15:56Z) - Neural-Symbolic Recursive Machine for Systematic Generalization [113.22455566135757]
We introduce the Neural-Symbolic Recursive Machine (NSR), whose core is a Grounded Symbol System (GSS)
NSR integrates neural perception, syntactic parsing, and semantic reasoning.
We evaluate NSR's efficacy across four challenging benchmarks designed to probe systematic generalization capabilities.
arXiv Detail & Related papers (2022-10-04T13:27:38Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Over-and-Under Complete Convolutional RNN for MRI Reconstruction [57.95363471940937]
Recent deep learning-based methods for MR image reconstruction usually leverage a generic auto-encoder architecture.
We propose an Over-and-Under Complete Convolu?tional Recurrent Neural Network (OUCR), which consists of an overcomplete and an undercomplete Convolutional Recurrent Neural Network(CRNN)
The proposed method achieves significant improvements over the compressed sensing and popular deep learning-based methods with less number of trainable parameters.
arXiv Detail & Related papers (2021-06-16T15:56:34Z) - Can RNNs learn Recursive Nested Subject-Verb Agreements? [4.094098809740732]
Language processing requires the ability to extract nested tree structures.
Recent advances in Recurrent Neural Networks (RNNs) achieve near-human performance in some language tasks.
arXiv Detail & Related papers (2021-01-06T20:47:02Z) - Incremental Training of a Recurrent Neural Network Exploiting a
Multi-Scale Dynamic Memory [79.42778415729475]
We propose a novel incrementally trained recurrent architecture targeting explicitly multi-scale learning.
We show how to extend the architecture of a simple RNN by separating its hidden state into different modules.
We discuss a training algorithm where new modules are iteratively added to the model to learn progressively longer dependencies.
arXiv Detail & Related papers (2020-06-29T08:35:49Z) - Separation of Memory and Processing in Dual Recurrent Neural Networks [0.0]
We explore a neural network architecture that stacks a recurrent layer and a feedforward layer that is also connected to the input.
When noise is introduced into the activation function of the recurrent units, these neurons are forced into a binary activation regime that makes the networks behave much as finite automata.
arXiv Detail & Related papers (2020-05-17T11:38:42Z) - How much complexity does an RNN architecture need to learn
syntax-sensitive dependencies? [9.248882589228089]
Long short-term memory (LSTM) networks are capable of encapsulating long-range dependencies.
Simple recurrent networks (SRNs) have generally been less successful at capturing long-range dependencies.
We propose a new architecture, the Decay RNN, which incorporates the decaying nature of neuronal activations.
arXiv Detail & Related papers (2020-05-17T09:13:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.