Fostering Event Compression using Gated Surprise
- URL: http://arxiv.org/abs/2005.05704v1
- Date: Tue, 12 May 2020 11:57:46 GMT
- Title: Fostering Event Compression using Gated Surprise
- Authors: Dania Humaidan, Sebastian Otte, Martin V. Butz
- Abstract summary: generative, event-predictive models are formed by segmenting sensorimotor data into chunks of contextual experiences.
Here, we introduce a hierarchical, surprise-gated recurrent neural network architecture, which models this process.
Our model shows to develop distinct event compressions and achieves the best performance on multiple event processing tasks.
- Score: 0.5801044612920815
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Our brain receives a dynamically changing stream of sensorimotor data. Yet,
we perceive a rather organized world, which we segment into and perceive as
events. Computational theories of cognitive science on event-predictive
cognition suggest that our brain forms generative, event-predictive models by
segmenting sensorimotor data into suitable chunks of contextual experiences.
Here, we introduce a hierarchical, surprise-gated recurrent neural network
architecture, which models this process and develops compact compressions of
distinct event-like contexts. The architecture contains a contextual LSTM
layer, which develops generative compressions of ongoing and subsequent
contexts. These compressions are passed into a GRU-like layer, which uses
surprise signals to update its recurrent latent state. The latent state is
passed forward into another LSTM layer, which processes actual dynamic sensory
flow in the light of the provided latent, contextual compression signals. Our
model shows to develop distinct event compressions and achieves the best
performance on multiple event processing tasks. The architecture may be very
useful for the further development of resource-efficient learning, hierarchical
model-based reinforcement learning, as well as the development of artificial
event-predictive cognition and intelligence.
Related papers
- Artificial Kuramoto Oscillatory Neurons [65.16453738828672]
We introduce Artificial Kuramotoy Neurons (AKOrN) as a dynamical alternative to threshold units.
We show that this idea provides performance improvements across a wide spectrum of tasks.
We believe that these empirical results show the importance of our assumptions at the most basic neuronal level of neural representation.
arXiv Detail & Related papers (2024-10-17T17:47:54Z) - Interpretable Spatio-Temporal Embedding for Brain Structural-Effective Network with Ordinary Differential Equation [56.34634121544929]
In this study, we first construct the brain-effective network via the dynamic causal model.
We then introduce an interpretable graph learning framework termed Spatio-Temporal Embedding ODE (STE-ODE)
This framework incorporates specifically designed directed node embedding layers, aiming at capturing the dynamic interplay between structural and effective networks.
arXiv Detail & Related papers (2024-05-21T20:37:07Z) - A developmental approach for training deep belief networks [0.46699574490885926]
Deep belief networks (DBNs) are neural networks that can extract rich internal representations of the environment from the sensory data.
We present iDBN, an iterative learning algorithm for DBNs that allows to jointly update the connection weights across all layers of the hierarchy.
Our work paves the way to the use of iDBN for modeling neurocognitive development.
arXiv Detail & Related papers (2022-07-12T11:37:58Z) - Developing hierarchical anticipations via neural network-based event
segmentation [14.059479351946386]
We model the development of hierarchical predictions via autonomously learned latent event codes.
We present a hierarchical recurrent neural network architecture, whose inductive learning biases foster the development of sparsely changing latent state.
A higher level network learns to predict the situations in which the latent states tend to change.
arXiv Detail & Related papers (2022-06-04T18:54:31Z) - CogNGen: Constructing the Kernel of a Hyperdimensional Predictive
Processing Cognitive Architecture [79.07468367923619]
We present a new cognitive architecture that combines two neurobiologically plausible, computational models.
We aim to develop a cognitive architecture that has the power of modern machine learning techniques.
arXiv Detail & Related papers (2022-03-31T04:44:28Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - Towards a Predictive Processing Implementation of the Common Model of
Cognition [79.63867412771461]
We describe an implementation of the common model of cognition grounded in neural generative coding and holographic associative memory.
The proposed system creates the groundwork for developing agents that learn continually from diverse tasks as well as model human performance at larger scales.
arXiv Detail & Related papers (2021-05-15T22:55:23Z) - Latent Event-Predictive Encodings through Counterfactual Regularization [0.9449650062296823]
We introduce a SUrprise-GAted Recurrent neural network (SUGAR) using a novel form of counterfactual regularization.
We test the model on a hierarchical sequence prediction task, where sequences are generated by alternating hidden graph structures.
arXiv Detail & Related papers (2021-05-12T18:30:09Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.