Learning stochastic dynamics and predicting emergent behavior using
transformers
- URL: http://arxiv.org/abs/2202.08708v1
- Date: Thu, 17 Feb 2022 15:27:21 GMT
- Title: Learning stochastic dynamics and predicting emergent behavior using
transformers
- Authors: Corneel Casert, Isaac Tamblyn and Stephen Whitelam
- Abstract summary: We show that a neural network can learn the dynamical rules of a system by observation of a single dynamical trajectory of the system.
We train a neural network called a transformer on a single trajectory of the model.
Transformers have the flexibility to learn dynamical rules from observation without explicit enumeration of rates or coarse-graining of configuration space.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We show that a neural network originally designed for language processing can
learn the dynamical rules of a stochastic system by observation of a single
dynamical trajectory of the system, and can accurately predict its emergent
behavior under conditions not observed during training. We consider a lattice
model of active matter undergoing continuous-time Monte Carlo dynamics,
simulated at a density at which its steady state comprises small, dispersed
clusters. We train a neural network called a transformer on a single trajectory
of the model. The transformer, which we show has the capacity to represent
dynamical rules that are numerous and nonlocal, learns that the dynamics of
this model consists of a small number of processes. Forward-propagated
trajectories of the trained transformer, at densities not encountered during
training, exhibit motility-induced phase separation and so predict the
existence of a nonequilibrium phase transition. Transformers have the
flexibility to learn dynamical rules from observation without explicit
enumeration of rates or coarse-graining of configuration space, and so the
procedure used here can be applied to a wide range of physical systems,
including those with large and complex dynamical generators.
Related papers
- Identification of Mean-Field Dynamics using Transformers [3.8916312075738273]
This paper investigates the use of transformer architectures to approximate the mean-field dynamics of particle systems exhibiting collective behavior.
Specifically, we prove that if a finite-dimensional transformer can effectively approximate the finite-dimensional vector field governing the particle system, then the expected output of this transformer provides a good approximation for the infinite-dimensional mean-field vector field.
arXiv Detail & Related papers (2024-10-06T19:47:24Z) - Latent Space Energy-based Neural ODEs [73.01344439786524]
This paper introduces a novel family of deep dynamical models designed to represent continuous-time sequence data.
We train the model using maximum likelihood estimation with Markov chain Monte Carlo.
Experiments on oscillating systems, videos and real-world state sequences (MuJoCo) illustrate that ODEs with the learnable energy-based prior outperform existing counterparts.
arXiv Detail & Related papers (2024-09-05T18:14:22Z) - Dynamical Mean-Field Theory of Self-Attention Neural Networks [0.0]
Transformer-based models have demonstrated exceptional performance across diverse domains.
Little is known about how they operate or what are their expected dynamics.
We use methods for the study of asymmetric Hopfield networks in nonequilibrium regimes.
arXiv Detail & Related papers (2024-06-11T13:29:34Z) - Latent State Models of Training Dynamics [51.88132043461152]
We train models with different random seeds and compute a variety of metrics throughout training.
We then fit a hidden Markov model (HMM) over the resulting sequences of metrics.
We use the HMM representation to study phase transitions and identify latent "detour" states that slow down convergence.
arXiv Detail & Related papers (2023-08-18T13:20:08Z) - Continuous Spatiotemporal Transformers [2.485182034310304]
We present the Continuous Stemporal Transformer (CST), a new transformer architecture that is designed to modeling continuous systems.
This new framework guarantees a continuous representation and output via optimization in Sobolev space.
We benchmark CST against traditional transformers as well as other smoothtemporal dynamics modeling methods and achieve superior performance in a number of tasks on synthetic and real systems.
arXiv Detail & Related papers (2023-01-31T00:06:56Z) - Physics-Inspired Temporal Learning of Quadrotor Dynamics for Accurate
Model Predictive Trajectory Tracking [76.27433308688592]
Accurately modeling quadrotor's system dynamics is critical for guaranteeing agile, safe, and stable navigation.
We present a novel Physics-Inspired Temporal Convolutional Network (PI-TCN) approach to learning quadrotor's system dynamics purely from robot experience.
Our approach combines the expressive power of sparse temporal convolutions and dense feed-forward connections to make accurate system predictions.
arXiv Detail & Related papers (2022-06-07T13:51:35Z) - Decomposed Linear Dynamical Systems (dLDS) for learning the latent
components of neural dynamics [6.829711787905569]
We propose a new decomposed dynamical system model that represents complex non-stationary and nonlinear dynamics of time series data.
Our model is trained through a dictionary learning procedure, where we leverage recent results in tracking sparse vectors over time.
In both continuous-time and discrete-time instructional examples we demonstrate that our model can well approximate the original system.
arXiv Detail & Related papers (2022-06-07T02:25:38Z) - Likelihood-Free Inference in State-Space Models with Unknown Dynamics [71.94716503075645]
We introduce a method for inferring and predicting latent states in state-space models where observations can only be simulated, and transition dynamics are unknown.
We propose a way of doing likelihood-free inference (LFI) of states and state prediction with a limited number of simulations.
arXiv Detail & Related papers (2021-11-02T12:33:42Z) - GEM: Group Enhanced Model for Learning Dynamical Control Systems [78.56159072162103]
We build effective dynamical models that are amenable to sample-based learning.
We show that learning the dynamics on a Lie algebra vector space is more effective than learning a direct state transition model.
This work sheds light on a connection between learning of dynamics and Lie group properties, which opens doors for new research directions.
arXiv Detail & Related papers (2021-04-07T01:08:18Z) - Euclideanizing Flows: Diffeomorphic Reduction for Learning Stable
Dynamical Systems [74.80320120264459]
We present an approach to learn such motions from a limited number of human demonstrations.
The complex motions are encoded as rollouts of a stable dynamical system.
The efficacy of this approach is demonstrated through validation on an established benchmark as well demonstrations collected on a real-world robotic system.
arXiv Detail & Related papers (2020-05-27T03:51:57Z) - Self-Supervised Learning of Generative Spin-Glasses with Normalizing
Flows [0.0]
We develop continuous spin-glass distributions with normalizing flows to model correlations in generic discrete problems.
We demonstrate that key physical and computational properties of the spin-glass phase can be successfully learned.
Remarkably, we observe that the learning itself corresponds to a spin-glass phase transition within the layers of the trained normalizing flows.
arXiv Detail & Related papers (2020-01-02T19:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.