Spatiotemporal Attention for Multivariate Time Series Prediction and
Interpretation
- URL: http://arxiv.org/abs/2008.04882v2
- Date: Mon, 26 Oct 2020 17:32:06 GMT
- Title: Spatiotemporal Attention for Multivariate Time Series Prediction and
Interpretation
- Authors: Tryambak Gangopadhyay, Sin Yong Tan, Zhanhong Jiang, Rui Meng, Soumik
Sarkar
- Abstract summary: temporal attention mechanism (STAM) for simultaneous learning of the most important time steps and variables.
Results: STAM maintains state-of-the-art prediction accuracy while offering the benefit of accurate interpretability.
- Score: 17.568599402858037
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multivariate time series modeling and prediction problems are abundant in
many machine learning application domains. Accurate interpretation of such
prediction outcomes from a machine learning model that explicitly captures
temporal correlations can significantly benefit the domain experts. In this
context, temporal attention has been successfully applied to isolate the
important time steps for the input time series. However, in multivariate time
series problems, spatial interpretation is also critical to understand the
contributions of different variables on the model outputs. We propose a novel
deep learning architecture, called spatiotemporal attention mechanism (STAM)
for simultaneous learning of the most important time steps and variables. STAM
is a causal (i.e., only depends on past inputs and does not use future inputs)
and scalable (i.e., scales well with an increase in the number of variables)
approach that is comparable to the state-of-the-art models in terms of
computational tractability. We demonstrate our models' performance on two
popular public datasets and a domain-specific dataset. When compared with the
baseline models, the results show that STAM maintains state-of-the-art
prediction accuracy while offering the benefit of accurate spatiotemporal
interpretability. The learned attention weights are validated from a domain
knowledge perspective for these real-world datasets.
Related papers
- DLFormer: Enhancing Explainability in Multivariate Time Series Forecasting using Distributed Lag Embedding [4.995397953581609]
This study introduces DLFormer, an attention-based architecture integrated with distributed lag embedding.
It showcases superior performance improvements compared to existing attention-based high-performance models.
arXiv Detail & Related papers (2024-08-29T20:39:54Z) - PDETime: Rethinking Long-Term Multivariate Time Series Forecasting from
the perspective of partial differential equations [49.80959046861793]
We present PDETime, a novel LMTF model inspired by the principles of Neural PDE solvers.
Our experimentation across seven diversetemporal real-world LMTF datasets reveals that PDETime adapts effectively to the intrinsic nature of the data.
arXiv Detail & Related papers (2024-02-25T17:39:44Z) - Graph Spatiotemporal Process for Multivariate Time Series Anomaly
Detection with Missing Values [67.76168547245237]
We introduce a novel framework called GST-Pro, which utilizes a graphtemporal process and anomaly scorer to detect anomalies.
Our experimental results show that the GST-Pro method can effectively detect anomalies in time series data and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2024-01-11T10:10:16Z) - GATGPT: A Pre-trained Large Language Model with Graph Attention Network
for Spatiotemporal Imputation [19.371155159744934]
In real-world settings, such data often contain missing elements due to issues like sensor malfunctions and data transmission errors.
The objective oftemporal imputation is to estimate these missing values by understanding the inherent spatial and temporal relationships in the observed time series.
Traditionally, intricatetemporal imputation has relied on specific architectures, which suffer from limited applicability and high computational complexity.
In contrast our approach integrates pre-trained large language models (LLMs) into intricatetemporal imputation, introducing a groundbreaking framework, GATGPT.
arXiv Detail & Related papers (2023-11-24T08:15:11Z) - An End-to-End Time Series Model for Simultaneous Imputation and Forecast [14.756607742477252]
We develop an end-to-end time series model that aims to learn the inference relation and make a multiple-step ahead forecast.
Our framework trains jointly two neural networks, one to learn the feature-wise correlations and the other for the modeling of temporal behaviors.
arXiv Detail & Related papers (2023-06-01T15:08:22Z) - Multi-scale Attention Flow for Probabilistic Time Series Forecasting [68.20798558048678]
We propose a novel non-autoregressive deep learning model, called Multi-scale Attention Normalizing Flow(MANF)
Our model avoids the influence of cumulative error and does not increase the time complexity.
Our model achieves state-of-the-art performance on many popular multivariate datasets.
arXiv Detail & Related papers (2022-05-16T07:53:42Z) - TACTiS: Transformer-Attentional Copulas for Time Series [76.71406465526454]
estimation of time-varying quantities is a fundamental component of decision making in fields such as healthcare and finance.
We propose a versatile method that estimates joint distributions using an attention-based decoder.
We show that our model produces state-of-the-art predictions on several real-world datasets.
arXiv Detail & Related papers (2022-02-07T21:37:29Z) - Model-Attentive Ensemble Learning for Sequence Modeling [86.4785354333566]
We present Model-Attentive Ensemble learning for Sequence modeling (MAES)
MAES is a mixture of time-series experts which leverages an attention-based gating mechanism to specialize the experts on different sequence dynamics and adaptively weight their predictions.
We demonstrate that MAES significantly out-performs popular sequence models on datasets subject to temporal shift.
arXiv Detail & Related papers (2021-02-23T05:23:35Z) - Transformer Hawkes Process [79.16290557505211]
We propose a Transformer Hawkes Process (THP) model, which leverages the self-attention mechanism to capture long-term dependencies.
THP outperforms existing models in terms of both likelihood and event prediction accuracy by a notable margin.
We provide a concrete example, where THP achieves improved prediction performance for learning multiple point processes when incorporating their relational information.
arXiv Detail & Related papers (2020-02-21T13:48:13Z) - Multivariate Probabilistic Time Series Forecasting via Conditioned
Normalizing Flows [8.859284959951204]
Time series forecasting is fundamental to scientific and engineering problems.
Deep learning methods are well suited for this problem.
We show that it improves over the state-of-the-art for standard metrics on many real-world data sets.
arXiv Detail & Related papers (2020-02-14T16:16:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.