Effectively Modeling Time Series with Simple Discrete State Spaces
- URL: http://arxiv.org/abs/2303.09489v1
- Date: Thu, 16 Mar 2023 17:08:21 GMT
- Title: Effectively Modeling Time Series with Simple Discrete State Spaces
- Authors: Michael Zhang, Khaled K. Saab, Michael Poli, Tri Dao, Karan Goel, and
Christopher R\'e
- Abstract summary: We introduce SpaceTime, a new state-space time series architecture for time series modeling.
For expressivity, we propose a new SSM parameterization based on the companion matrix.
For long horizon forecasting, we introduce a "closed-loop" variation of the companion SSM.
For efficient training and inference, we introduce an algorithm that reduces the memory and compute of a forward pass with the companion matrix.
- Score: 25.046268609133982
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Time series modeling is a well-established problem, which often requires that
methods (1) expressively represent complicated dependencies, (2) forecast long
horizons, and (3) efficiently train over long sequences. State-space models
(SSMs) are classical models for time series, and prior works combine SSMs with
deep learning layers for efficient sequence modeling. However, we find
fundamental limitations with these prior approaches, proving their SSM
representations cannot express autoregressive time series processes. We thus
introduce SpaceTime, a new state-space time series architecture that improves
all three criteria. For expressivity, we propose a new SSM parameterization
based on the companion matrix -- a canonical representation for discrete-time
processes -- which enables SpaceTime's SSM layers to learn desirable
autoregressive processes. For long horizon forecasting, we introduce a
"closed-loop" variation of the companion SSM, which enables SpaceTime to
predict many future time-steps by generating its own layer-wise inputs. For
efficient training and inference, we introduce an algorithm that reduces the
memory and compute of a forward pass with the companion matrix. With sequence
length $\ell$ and state-space size $d$, we go from $\tilde{O}(d \ell)$
na\"ively to $\tilde{O}(d + \ell)$. In experiments, our contributions lead to
state-of-the-art results on extensive and diverse benchmarks, with best or
second-best AUROC on 6 / 7 ECG and speech time series classification, and best
MSE on 14 / 16 Informer forecasting tasks. Furthermore, we find SpaceTime (1)
fits AR($p$) processes that prior deep SSMs fail on, (2) forecasts notably more
accurately on longer horizons than prior state-of-the-art, and (3) speeds up
training on real-world ETTh1 data by 73% and 80% relative wall-clock time over
Transformers and LSTMs.
Related papers
- Coarse Correspondences Boost Spatial-Temporal Reasoning in Multimodal Language Model [51.83436609094658]
We introduce Coarse Correspondences, a simple lightweight method that enhances MLLMs' spatial-temporal reasoning with 2D images as input.
Our method uses a lightweight tracking model to identify primary object correspondences between frames in a video or across different image viewpoints.
We demonstrate that this simple training-free approach brings substantial gains to GPT4-V/O consistently on four benchmarks.
arXiv Detail & Related papers (2024-08-01T17:57:12Z) - Deciphering Movement: Unified Trajectory Generation Model for Multi-Agent [53.637837706712794]
We propose a Unified Trajectory Generation model, UniTraj, that processes arbitrary trajectories as masked inputs.
Specifically, we introduce a Ghost Spatial Masking (GSM) module embedded within a Transformer encoder for spatial feature extraction.
We benchmark three practical sports game datasets, Basketball-U, Football-U, and Soccer-U, for evaluation.
arXiv Detail & Related papers (2024-05-27T22:15:23Z) - Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting [22.84798547604491]
State Space Models (SSMs) approximate continuous systems using a set of basis functions and discretize them to handle input data.
This paper proposes a novel theoretical framework termed Dynamic Spectral Operator, offering more intuitive and general guidance on applying SSMs to time series data.
We introduce Time-SSM, a novel SSM-based foundation model with only one-seventh of the parameters compared to Mamba.
arXiv Detail & Related papers (2024-05-25T17:42:40Z) - Convolutional State Space Models for Long-Range Spatiotemporal Modeling [65.0993000439043]
ConvS5 is an efficient variant for long-rangetemporal modeling.
It significantly outperforms Transformers and ConvNISTTM on a long horizon Moving-Lab experiment while training 3X faster than ConvLSTM and generating samples 400X faster than Transformers.
arXiv Detail & Related papers (2023-10-30T16:11:06Z) - FormerTime: Hierarchical Multi-Scale Representations for Multivariate
Time Series Classification [53.55504611255664]
FormerTime is a hierarchical representation model for improving the classification capacity for the multivariate time series classification task.
It exhibits three aspects of merits: (1) learning hierarchical multi-scale representations from time series data, (2) inheriting the strength of both transformers and convolutional networks, and (3) tacking the efficiency challenges incurred by the self-attention mechanism.
arXiv Detail & Related papers (2023-02-20T07:46:14Z) - P-STMO: Pre-Trained Spatial Temporal Many-to-One Model for 3D Human Pose
Estimation [78.83305967085413]
This paper introduces a novel Pre-trained Spatial Temporal Many-to-One (P-STMO) model for 2D-to-3D human pose estimation task.
Our method outperforms state-of-the-art methods with fewer parameters and less computational overhead.
arXiv Detail & Related papers (2022-03-15T04:00:59Z) - Efficiently Modeling Long Sequences with Structured State Spaces [15.456254157293836]
We propose a new sequence model based on a new parameterization for the fundamental state space model.
S4 achieves strong empirical results across a diverse range of established benchmarks, including (i) 91% accuracy on sequential CIFAR-10 with no data augmentation or auxiliary losses, on par with a larger 2-D ResNet.
arXiv Detail & Related papers (2021-10-31T03:32:18Z) - Combining Recurrent, Convolutional, and Continuous-time Models with
Linear State-Space Layers [21.09321438439848]
We introduce a simple sequence model inspired by control systems that generalize.
We show that LSSL models are closely related to the three aforementioned families of models and inherit their strengths.
For example, they generalize convolutions to continuous-time, explain common RNN-1s, and share features of NDEs such as time-scale adaptation.
arXiv Detail & Related papers (2021-10-26T19:44:53Z) - Convolutional Tensor-Train LSTM for Spatio-temporal Learning [116.24172387469994]
We propose a higher-order LSTM model that can efficiently learn long-term correlations in the video sequence.
This is accomplished through a novel tensor train module that performs prediction by combining convolutional features across time.
Our results achieve state-of-the-art performance-art in a wide range of applications and datasets.
arXiv Detail & Related papers (2020-02-21T05:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.