MLPST: MLP is All You Need for Spatio-Temporal Prediction
- URL: http://arxiv.org/abs/2309.13363v1
- Date: Sat, 23 Sep 2023 12:58:16 GMT
- Title: MLPST: MLP is All You Need for Spatio-Temporal Prediction
- Authors: Zijian Zhang, Ze Huang, Zhiwei Hu, Xiangyu Zhao, Wanyu Wang, Zitao
Liu, Junbo Zhang, S. Joe Qin and Hongwei Zhao
- Abstract summary: Traffic is a typical deep model-temporal-based prediction method.
We propose a pure multi-layer perceptron architecture for traffic prediction.
- Score: 40.65579041549435
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traffic prediction is a typical spatio-temporal data mining task and has
great significance to the public transportation system. Considering the demand
for its grand application, we recognize key factors for an ideal
spatio-temporal prediction method: efficient, lightweight, and effective.
However, the current deep model-based spatio-temporal prediction solutions
generally own intricate architectures with cumbersome optimization, which can
hardly meet these expectations. To accomplish the above goals, we propose an
intuitive and novel framework, MLPST, a pure multi-layer perceptron
architecture for traffic prediction. Specifically, we first capture spatial
relationships from both local and global receptive fields. Then, temporal
dependencies in different intervals are comprehensively considered. Through
compact and swift MLP processing, MLPST can well capture the spatial and
temporal dependencies while requiring only linear computational complexity, as
well as model parameters that are more than an order of magnitude lower than
baselines. Extensive experiments validated the superior effectiveness and
efficiency of MLPST against advanced baselines, and among models with optimal
accuracy, MLPST achieves the best time and space efficiency.
Related papers
- ST-Mamba: Spatial-Temporal Selective State Space Model for Traffic Flow Prediction [32.44888387725925]
The proposed ST-Mamba model is first to leverage the power of spatial-temporal learning in traffic flow prediction without using graph modeling.
The proposed ST-Mamba model achieves a 61.11% improvement in computational speed and increases prediction accuracy by 0.67%.
Experiments with real-world traffic datasets demonstrate that the textsfST-Mamba model sets a new benchmark in traffic flow prediction.
arXiv Detail & Related papers (2024-04-20T03:57:57Z) - AMP: Autoregressive Motion Prediction Revisited with Next Token Prediction for Autonomous Driving [59.94343412438211]
We introduce the GPT style next token motion prediction into motion prediction.
Different from language data which is composed of homogeneous units -words, the elements in the driving scene could have complex spatial-temporal and semantic relations.
We propose to adopt three factorized attention modules with different neighbors for information aggregation and different position encoding styles to capture their relations.
arXiv Detail & Related papers (2024-03-20T06:22:37Z) - SIMPL: A Simple and Efficient Multi-agent Motion Prediction Baseline for
Autonomous Driving [27.776472262857045]
This paper presents a Simple and effIcient Motion Prediction baseLine (SIMPL) for autonomous vehicles.
We propose a compact and efficient global feature fusion module that performs directed message passing in a symmetric manner.
As a strong baseline, SIMPL exhibits highly competitive performance on Argoverse 1 & 2 motion forecasting benchmarks.
arXiv Detail & Related papers (2024-02-04T15:07:49Z) - Spatial-Temporal Large Language Model for Traffic Prediction [21.69991612610926]
We propose a Spatial-Temporal Large Language Model (ST-LLM) for traffic prediction.
In the ST-LLM, we define timesteps at each location as tokens and design a spatial-temporal embedding to learn the spatial location and global temporal patterns of these tokens.
In experiments on real traffic datasets, ST-LLM is a powerful spatial-temporal learner that outperforms state-of-the-art models.
arXiv Detail & Related papers (2024-01-18T17:03:59Z) - Frequency-domain MLPs are More Effective Learners in Time Series
Forecasting [67.60443290781988]
Time series forecasting has played the key role in different industrial domains, including finance, traffic, energy, and healthcare.
Most-based forecasting methods suffer from the point-wise mappings and information bottleneck.
We propose FreTS, a simple yet effective architecture built upon Frequency-domains for Time Series forecasting.
arXiv Detail & Related papers (2023-11-10T17:05:13Z) - ST-MLP: A Cascaded Spatio-Temporal Linear Framework with
Channel-Independence Strategy for Traffic Forecasting [47.74479442786052]
Current research on Spatio-Temporal Graph Neural Networks (STGNNs) often prioritizes complex designs, leading to computational burdens with only minor enhancements in accuracy.
We propose ST-MLP, a concise cascaded temporal-temporal model solely based on Multi-Layer Perceptron (MLP) modules and linear layers.
Empirical results demonstrate that ST-MLP outperforms state-of-the-art STGNNs and other models in terms of accuracy and computational efficiency.
arXiv Detail & Related papers (2023-08-14T23:34:59Z) - Contextualizing MLP-Mixers Spatiotemporally for Urban Data Forecast at
Scale [57.38373754100004]
Stemporal urban data (STUD) displays complex correlational patterns.
Because STUD is often massive in scale, practitioners need to strike a balance between effectiveness and efficiency.
An alternative paradigm called Nex-Mixer has the potential for both simplicity and effectiveness.
arXiv Detail & Related papers (2023-07-04T05:19:19Z) - COST-EFF: Collaborative Optimization of Spatial and Temporal Efficiency
with Slenderized Multi-exit Language Models [16.586312156966635]
Transformer-based pre-trained language models (PLMs) mostly suffer from excessive overhead despite their advanced capacity.
Existing statically compressed models are unaware of the diverse complexities between input instances.
We propose a collaborative optimization for PLMs that integrates static model compression and dynamic inference acceleration.
arXiv Detail & Related papers (2022-10-27T15:06:40Z) - Learning representations with end-to-end models for improved remaining
useful life prognostics [64.80885001058572]
The remaining Useful Life (RUL) of equipment is defined as the duration between the current time and its failure.
We propose an end-to-end deep learning model based on multi-layer perceptron and long short-term memory layers (LSTM) to predict the RUL.
We will discuss how the proposed end-to-end model is able to achieve such good results and compare it to other deep learning and state-of-the-art methods.
arXiv Detail & Related papers (2021-04-11T16:45:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.