Latent Event-Predictive Encodings through Counterfactual Regularization
- URL: http://arxiv.org/abs/2105.05894v1
- Date: Wed, 12 May 2021 18:30:09 GMT
- Title: Latent Event-Predictive Encodings through Counterfactual Regularization
- Authors: Dania Humaidan, Sebastian Otte, Christian Gumbsch, Charley Wu, Martin
V. Butz
- Abstract summary: We introduce a SUrprise-GAted Recurrent neural network (SUGAR) using a novel form of counterfactual regularization.
We test the model on a hierarchical sequence prediction task, where sequences are generated by alternating hidden graph structures.
- Score: 0.9449650062296823
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A critical challenge for any intelligent system is to infer structure from
continuous data streams. Theories of event-predictive cognition suggest that
the brain segments sensorimotor information into compact event encodings, which
are used to anticipate and interpret environmental dynamics. Here, we introduce
a SUrprise-GAted Recurrent neural network (SUGAR) using a novel form of
counterfactual regularization. We test the model on a hierarchical sequence
prediction task, where sequences are generated by alternating hidden graph
structures. Our model learns to both compress the temporal dynamics of the task
into latent event-predictive encodings and anticipate event transitions at the
right moments, given noisy hidden signals about them. The addition of the
counterfactual regularization term ensures fluid transitions from one latent
code to the next, whereby the resulting latent codes exhibit compositional
properties. The implemented mechanisms offer a host of useful applications in
other domains, including hierarchical reasoning, planning, and decision making.
Related papers
- Trajectory Forecasting through Low-Rank Adaptation of Discrete Latent Codes [36.12653178844828]
Trajectory forecasting is crucial for video surveillance analytics, as it enables the anticipation of future movements for a set of agents.
We introduce Vector Quantized Variational Autoencoders (VQ-VAEs), which utilize a discrete latent space to tackle the issue of posterior collapse.
We show that such a two-fold framework, augmented with instance-level discretization, leads to accurate and diverse forecasts.
arXiv Detail & Related papers (2024-05-31T10:13:17Z) - Uncovering the Missing Pattern: Unified Framework Towards Trajectory
Imputation and Prediction [60.60223171143206]
Trajectory prediction is a crucial undertaking in understanding entity movement or human behavior from observed sequences.
Current methods often assume that the observed sequences are complete while ignoring the potential for missing values.
This paper presents a unified framework, the Graph-based Conditional Variational Recurrent Neural Network (GC-VRNN), which can perform trajectory imputation and prediction simultaneously.
arXiv Detail & Related papers (2023-03-28T14:27:27Z) - Developing hierarchical anticipations via neural network-based event
segmentation [14.059479351946386]
We model the development of hierarchical predictions via autonomously learned latent event codes.
We present a hierarchical recurrent neural network architecture, whose inductive learning biases foster the development of sparsely changing latent state.
A higher level network learns to predict the situations in which the latent states tend to change.
arXiv Detail & Related papers (2022-06-04T18:54:31Z) - Hybrid Predictive Coding: Inferring, Fast and Slow [62.997667081978825]
We propose a hybrid predictive coding network that combines both iterative and amortized inference in a principled manner.
We demonstrate that our model is inherently sensitive to its uncertainty and adaptively balances balances to obtain accurate beliefs using minimum computational expense.
arXiv Detail & Related papers (2022-04-05T12:52:45Z) - Differentiable Generalised Predictive Coding [2.868176771215219]
This paper deals with differentiable dynamical models congruent with neural process theories that cast brain function as the hierarchical refinement of an internal generative model explaining observations.
Our work extends existing implementations of gradient-based predictive coding and allows to integrate deep neural networks for non-linear state parameterization.
arXiv Detail & Related papers (2021-12-02T22:02:56Z) - DeepExpress: Heterogeneous and Coupled Sequence Modeling for Express
Delivery Prediction [41.49535437719119]
A precise estimate of consumer delivery requests has to involve sequential factors such as shopping behaviors, weather conditions, events, business campaigns, and their couplings.
DeepExpress is a deep-learning based express delivery sequence prediction model, which extends the classic seq2seq framework to learning complex coupling between sequence and features.
arXiv Detail & Related papers (2021-08-18T14:24:19Z) - CCVS: Context-aware Controllable Video Synthesis [95.22008742695772]
presentation introduces a self-supervised learning approach to the synthesis of new video clips from old ones.
It conditions the synthesis process on contextual information for temporal continuity and ancillary information for fine control.
arXiv Detail & Related papers (2021-07-16T17:57:44Z) - Temporal Predictive Coding For Model-Based Planning In Latent Space [80.99554006174093]
We present an information-theoretic approach that employs temporal predictive coding to encode elements in the environment that can be predicted across time.
We evaluate our model on a challenging modification of standard DMControl tasks where the background is replaced with natural videos that contain complex but irrelevant information to the planning task.
arXiv Detail & Related papers (2021-06-14T04:31:15Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Representation Learning for Sequence Data with Deep Autoencoding
Predictive Components [96.42805872177067]
We propose a self-supervised representation learning method for sequence data, based on the intuition that useful representations of sequence data should exhibit a simple structure in the latent space.
We encourage this latent structure by maximizing an estimate of predictive information of latent feature sequences, which is the mutual information between past and future windows at each time step.
We demonstrate that our method recovers the latent space of noisy dynamical systems, extracts predictive features for forecasting tasks, and improves automatic speech recognition when used to pretrain the encoder on large amounts of unlabeled data.
arXiv Detail & Related papers (2020-10-07T03:34:01Z) - Fostering Event Compression using Gated Surprise [0.5801044612920815]
generative, event-predictive models are formed by segmenting sensorimotor data into chunks of contextual experiences.
Here, we introduce a hierarchical, surprise-gated recurrent neural network architecture, which models this process.
Our model shows to develop distinct event compressions and achieves the best performance on multiple event processing tasks.
arXiv Detail & Related papers (2020-05-12T11:57:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.