Discovering group dynamics in synchronous time series via hierarchical
recurrent switching-state models
- URL: http://arxiv.org/abs/2401.14973v1
- Date: Fri, 26 Jan 2024 16:06:01 GMT
- Title: Discovering group dynamics in synchronous time series via hierarchical
recurrent switching-state models
- Authors: Michael Wojnowicz, Preetish Rath, Eric Miller, Jeffrey Miller,
Clifford Hancock, Meghan O'Donovan, Seth Elkin-Frankston, Thaddeus Brunye,
and Michael C. Hughes
- Abstract summary: We seek to model a collection of time series arising from multiple entities interacting over the same time period.
Recent work focused on modeling individual time series is inadequate for our intended applications, where collective system-level behavior influences the trajectories of individual entities.
We employ a latent system-level discrete state chain that drives latent entity-level chains which in turn govern the dynamics of each observed time series.
- Score: 3.5902825399592193
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We seek to model a collection of time series arising from multiple entities
interacting over the same time period. Recent work focused on modeling
individual time series is inadequate for our intended applications, where
collective system-level behavior influences the trajectories of individual
entities. To address such problems, we present a new hierarchical
switching-state model that can be trained in an unsupervised fashion to
simultaneously explain both system-level and individual-level dynamics. We
employ a latent system-level discrete state Markov chain that drives latent
entity-level chains which in turn govern the dynamics of each observed time
series. Feedback from the observations to the chains at both the entity and
system levels improves flexibility via context-dependent state transitions. Our
hierarchical switching recurrent dynamical models can be learned via
closed-form variational coordinate ascent updates to all latent chains that
scale linearly in the number of individual time series. This is asymptotically
no more costly than fitting separate models for each entity. Experiments on
synthetic and real datasets show that our model can produce better forecasts of
future entity behavior than existing methods. Moreover, the availability of
latent state chains at both the entity and system level enables interpretation
of group dynamics.
Related papers
- Enhancing deep neural networks through complex-valued representations and Kuramoto synchronization dynamics [13.611995923070426]
We investigate whether synchrony-based mechanisms can enhance object encoding in artificial models trained for visual categorization.
We combine complex-valued representations with Kuramoto dynamics to promote phase alignment, facilitating the grouping of features belonging to the same object.
Our findings highlight the potential of synchrony-driven mechanisms to enhance deep learning models, improving their performance, robustness, and generalization.
arXiv Detail & Related papers (2025-02-28T14:10:42Z) - Learning Interpretable Hierarchical Dynamical Systems Models from Time Series Data [6.3128614613706295]
We show how to efficiently harvest group-level (multi-domain) information while retaining single-domain dynamical characteristics.
In addition to faithful reconstruction of all individual dynamical regimes, our unsupervised methodology discovers common low-dimensional feature spaces.
arXiv Detail & Related papers (2024-10-07T07:54:53Z) - Dynamic Post-Hoc Neural Ensemblers [55.15643209328513]
In this study, we explore employing neural networks as ensemble methods.
Motivated by the risk of learning low-diversity ensembles, we propose regularizing the model by randomly dropping base model predictions.
We demonstrate this approach lower bounds the diversity within the ensemble, reducing overfitting and improving generalization capabilities.
arXiv Detail & Related papers (2024-10-06T15:25:39Z) - Latent Space Energy-based Neural ODEs [73.01344439786524]
This paper introduces a novel family of deep dynamical models designed to represent continuous-time sequence data.
We train the model using maximum likelihood estimation with Markov chain Monte Carlo.
Experiments on oscillating systems, videos and real-world state sequences (MuJoCo) illustrate that ODEs with the learnable energy-based prior outperform existing counterparts.
arXiv Detail & Related papers (2024-09-05T18:14:22Z) - Neural Persistence Dynamics [8.197801260302642]
We consider the problem of learning the dynamics in the topology of time-evolving point clouds.
Our proposed model - $textitNeural Persistence Dynamics$ - substantially outperforms the state-of-the-art across a diverse set of parameter regression tasks.
arXiv Detail & Related papers (2024-05-24T17:20:18Z) - Attractor Memory for Long-Term Time Series Forecasting: A Chaos Perspective [63.60312929416228]
textbftextitAttraos incorporates chaos theory into long-term time series forecasting.
We show that Attraos outperforms various LTSF methods on mainstream datasets and chaotic datasets with only one-twelfth of the parameters compared to PatchTST.
arXiv Detail & Related papers (2024-02-18T05:35:01Z) - TimeGraphs: Graph-based Temporal Reasoning [64.18083371645956]
TimeGraphs is a novel approach that characterizes dynamic interactions as a hierarchical temporal graph.
Our approach models the interactions using a compact graph-based representation, enabling adaptive reasoning across diverse time scales.
We evaluate TimeGraphs on multiple datasets with complex, dynamic agent interactions, including a football simulator, the Resistance game, and the MOMA human activity dataset.
arXiv Detail & Related papers (2024-01-06T06:26:49Z) - Decomposed Linear Dynamical Systems (dLDS) for learning the latent
components of neural dynamics [6.829711787905569]
We propose a new decomposed dynamical system model that represents complex non-stationary and nonlinear dynamics of time series data.
Our model is trained through a dictionary learning procedure, where we leverage recent results in tracking sparse vectors over time.
In both continuous-time and discrete-time instructional examples we demonstrate that our model can well approximate the original system.
arXiv Detail & Related papers (2022-06-07T02:25:38Z) - Deep Explicit Duration Switching Models for Time Series [84.33678003781908]
We propose a flexible model that is capable of identifying both state- and time-dependent switching dynamics.
State-dependent switching is enabled by a recurrent state-to-switch connection.
An explicit duration count variable is used to improve the time-dependent switching behavior.
arXiv Detail & Related papers (2021-10-26T17:35:21Z) - Variational Predictive Routing with Nested Subjective Timescales [1.6114012813668934]
We present Variational Predictive Routing (PRV) - a neural inference system that organizes latent video features in a temporal hierarchy.
We show that VPR is able to detect event boundaries, disentangletemporal features, adapt to the dynamics hierarchy of the data, and produce accurate time-agnostic rollouts of the future.
arXiv Detail & Related papers (2021-10-21T16:12:59Z) - Chaos as an interpretable benchmark for forecasting and data-driven
modelling [7.6146285961466]
Chaotic systems pose a unique challenge to modern statistical learning techniques.
We present a database currently comprising 131 known chaotic dynamical systems spanning fields such as astrophysics, climatology, and biochemistry.
arXiv Detail & Related papers (2021-10-11T13:39:41Z) - Modeling Regime Shifts in Multiple Time Series [1.4588552933974936]
Regime shifts refer to the changing behaviors exhibited by series at different time intervals.
Existing methods fail to take relationships between time series into consideration for discovering regimes in multiple time series.
Most of the existing methods are unable to handle all of these three issues in a unified framework.
arXiv Detail & Related papers (2021-09-20T17:02:29Z) - TCL: Transformer-based Dynamic Graph Modelling via Contrastive Learning [87.38675639186405]
We propose a novel graph neural network approach, called TCL, which deals with the dynamically-evolving graph in a continuous-time fashion.
To the best of our knowledge, this is the first attempt to apply contrastive learning to representation learning on dynamic graphs.
arXiv Detail & Related papers (2021-05-17T15:33:25Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Learning Temporal Dynamics from Cycles in Narrated Video [85.89096034281694]
We propose a self-supervised solution to the problem of learning to model how the world changes as time elapses.
Our model learns modality-agnostic functions to predict forward and backward in time, which must undo each other when composed.
We apply the learned dynamics model without further training to various tasks, such as predicting future action and temporally ordering sets of images.
arXiv Detail & Related papers (2021-01-07T02:41:32Z) - Continuous-in-Depth Neural Networks [107.47887213490134]
We first show that ResNets fail to be meaningful dynamical in this richer sense.
We then demonstrate that neural network models can learn to represent continuous dynamical systems.
We introduce ContinuousNet as a continuous-in-depth generalization of ResNet architectures.
arXiv Detail & Related papers (2020-08-05T22:54:09Z) - S2RMs: Spatially Structured Recurrent Modules [105.0377129434636]
We take a step towards exploiting dynamic structure that are capable of simultaneously exploiting both modular andtemporal structures.
We find our models to be robust to the number of available views and better capable of generalization to novel tasks without additional training.
arXiv Detail & Related papers (2020-07-13T17:44:30Z) - Context-aware Dynamics Model for Generalization in Model-Based
Reinforcement Learning [124.9856253431878]
We decompose the task of learning a global dynamics model into two stages: (a) learning a context latent vector that captures the local dynamics, then (b) predicting the next state conditioned on it.
In order to encode dynamics-specific information into the context latent vector, we introduce a novel loss function that encourages the context latent vector to be useful for predicting both forward and backward dynamics.
The proposed method achieves superior generalization ability across various simulated robotics and control tasks, compared to existing RL schemes.
arXiv Detail & Related papers (2020-05-14T08:10:54Z) - Relational State-Space Model for Stochastic Multi-Object Systems [24.234120525358456]
This paper introduces the relational state-space model (R-SSM), a sequential hierarchical latent variable model.
R-SSM makes use of graph neural networks (GNNs) to simulate the joint state transitions of multiple correlated objects.
The utility of R-SSM is empirically evaluated on synthetic and real time-series datasets.
arXiv Detail & Related papers (2020-01-13T03:45:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.