Contextualizing MLP-Mixers Spatiotemporally for Urban Data Forecast at Scale
- URL: http://arxiv.org/abs/2307.01482v6
- Date: Sat, 02 Nov 2024 03:19:24 GMT
- Title: Contextualizing MLP-Mixers Spatiotemporally for Urban Data Forecast at Scale
- Authors: Tong Nie, Guoyang Qin, Lijun Sun, Wei Ma, Yu Mei, Jian Sun,
- Abstract summary: We propose an adapted version of the computationally-Mixer for STTD forecast at scale.
Our results surprisingly show that this simple-yeteffective solution can rival SOTA baselines when tested on several traffic benchmarks.
Our findings contribute to the exploration of simple-yet-effective models for real-world STTD forecasting.
- Score: 54.15522908057831
- License:
- Abstract: Spatiotemporal traffic data (STTD) displays complex correlational structures. Extensive advanced techniques have been designed to capture these structures for effective forecasting. However, because STTD is often massive in scale, practitioners need to strike a balance between effectiveness and efficiency using computationally efficient models. An alternative paradigm based on multilayer perceptron (MLP) called MLP-Mixer has the potential for both simplicity and effectiveness. Taking inspiration from its success in other domains, we propose an adapted version, named NexuSQN, for STTD forecast at scale. We first identify the challenges faced when directly applying MLP-Mixers as seriesand window-wise multivaluedness. To distinguish between spatial and temporal patterns, the concept of ST-contextualization is then proposed. Our results surprisingly show that this simple-yeteffective solution can rival SOTA baselines when tested on several traffic benchmarks. Furthermore, NexuSQN has demonstrated its versatility across different domains, including energy and environment data, and has been deployed in a collaborative project with Baidu to predict congestion in megacities like Beijing and Shanghai. Our findings contribute to the exploration of simple-yet-effective models for real-world STTD forecasting.
Related papers
- PreMixer: MLP-Based Pre-training Enhanced MLP-Mixers for Large-scale Traffic Forecasting [30.055634767677823]
In urban computing, precise and swift forecasting of time series data from traffic networks is crucial.
Current research limitations because of inherent inefficiency of model and their unsuitability for large-scale traffic applications due to model complexity.
This paper proposes a novel framework, named PreMixer, designed to bridge this gap. It features a predictive model and a pre-training mechanism, both based on the principles of Multi-Layer Perceptrons (MLP)
Our framework achieves comparable state-of-theart performance while maintaining high computational efficiency, as verified by extensive experiments on large-scale traffic datasets.
arXiv Detail & Related papers (2024-12-18T08:35:40Z) - Cross Space and Time: A Spatio-Temporal Unitized Model for Traffic Flow Forecasting [16.782154479264126]
Predicting backbone-temporal traffic flow presents challenges due to complex interactions between temporal factors.
Existing approaches address these dimensions in isolation, neglecting their critical interdependencies.
In this paper, we introduce Sanonymous-Temporal Unitized Unitized Cell (ASTUC), a unified framework designed to capture both spatial and temporal dependencies.
arXiv Detail & Related papers (2024-11-14T07:34:31Z) - Sparse Prototype Network for Explainable Pedestrian Behavior Prediction [60.80524827122901]
We present Sparse Prototype Network (SPN), an explainable method designed to simultaneously predict a pedestrian's future action, trajectory, and pose.
Regularized by mono-semanticity and clustering constraints, the prototypes learn consistent and human-understandable features.
arXiv Detail & Related papers (2024-10-16T03:33:40Z) - MITA: Bridging the Gap between Model and Data for Test-time Adaptation [68.62509948690698]
Test-Time Adaptation (TTA) has emerged as a promising paradigm for enhancing the generalizability of models.
We propose Meet-In-The-Middle based MITA, which introduces energy-based optimization to encourage mutual adaptation of the model and data from opposing directions.
arXiv Detail & Related papers (2024-10-12T07:02:33Z) - UniMix: Towards Domain Adaptive and Generalizable LiDAR Semantic Segmentation in Adverse Weather [55.95708988160047]
LiDAR semantic segmentation (LSS) is a critical task in autonomous driving.
Prior LSS methods are investigated and evaluated on datasets within the same domain in clear weather.
We propose UniMix, a universal method that enhances the adaptability and generalizability of LSS models.
arXiv Detail & Related papers (2024-04-08T02:02:15Z) - RPMixer: Shaking Up Time Series Forecasting with Random Projections for Large Spatial-Temporal Data [33.0546525587517]
We propose a all-Multi-Layer Perceptron (all-MLP) time series forecasting architecture called RPMixer.
Our method capitalizes on the ensemble-like behavior of deep neural networks, where each individual block behaves like a base learner in an ensemble model.
arXiv Detail & Related papers (2024-02-16T07:28:59Z) - MLPST: MLP is All You Need for Spatio-Temporal Prediction [40.65579041549435]
Traffic is a typical deep model-temporal-based prediction method.
We propose a pure multi-layer perceptron architecture for traffic prediction.
arXiv Detail & Related papers (2023-09-23T12:58:16Z) - Spatial-Temporal Identity: A Simple yet Effective Baseline for
Multivariate Time Series Forecasting [17.84296081495185]
We explore the critical factors of MTS forecasting and design a model that is as powerful as STGNNs, but more concise and efficient.
We identify the indistinguishability of samples in both spatial and temporal dimensions as a key bottleneck, and propose a simple yet effective baseline for MTS forecasting.
These results suggest that we can design efficient and effective models as long as they solve the indistinguishability of samples, without being limited to STGNNs.
arXiv Detail & Related papers (2022-08-10T09:25:43Z) - Distributionally Robust Models with Parametric Likelihood Ratios [123.05074253513935]
Three simple ideas allow us to train models with DRO using a broader class of parametric likelihood ratios.
We find that models trained with the resulting parametric adversaries are consistently more robust to subpopulation shifts when compared to other DRO approaches.
arXiv Detail & Related papers (2022-04-13T12:43:12Z) - A Generative Learning Approach for Spatio-temporal Modeling in Connected
Vehicular Network [55.852401381113786]
This paper proposes LaMI (Latency Model Inpainting), a novel framework to generate a comprehensive-temporal quality framework for wireless access latency of connected vehicles.
LaMI adopts the idea from image inpainting and synthesizing and can reconstruct the missing latency samples by a two-step procedure.
In particular, it first discovers the spatial correlation between samples collected in various regions using a patching-based approach and then feeds the original and highly correlated samples into a Varienational Autocoder (VAE)
arXiv Detail & Related papers (2020-03-16T03:43:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.