Hybrid hidden Markov LSTM for short-term traffic flow prediction
- URL: http://arxiv.org/abs/2307.04954v2
- Date: Mon, 17 Jul 2023 00:15:30 GMT
- Title: Hybrid hidden Markov LSTM for short-term traffic flow prediction
- Authors: Agnimitra Sengupta, Adway Das, S. Ilgin Guler
- Abstract summary: We propose a hybrid hidden Markov-LSTM model that is capable of learning complementary features in traffic data.
Results indicate significant performance gains in using hybrid architecture compared to conventional methods.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning (DL) methods have outperformed parametric models such as
historical average, ARIMA and variants in predicting traffic variables into
short and near-short future, that are critical for traffic management.
Specifically, recurrent neural network (RNN) and its variants (e.g. long
short-term memory) are designed to retain long-term temporal correlations and
therefore are suitable for modeling sequences. However, multi-regime models
assume the traffic system to evolve through multiple states (say, free-flow,
congestion in traffic) with distinct characteristics, and hence, separate
models are trained to characterize the traffic dynamics within each regime. For
instance, Markov-switching models with a hidden Markov model (HMM) for regime
identification is capable of capturing complex dynamic patterns and
non-stationarity. Interestingly, both HMM and LSTM can be used for modeling an
observation sequence from a set of latent or, hidden state variables. In LSTM,
the latent variable is computed in a deterministic manner from the current
observation and the previous latent variable, while, in HMM, the set of latent
variables is a Markov chain. Inspired by research in natural language
processing, a hybrid hidden Markov-LSTM model that is capable of learning
complementary features in traffic data is proposed for traffic flow prediction.
Results indicate significant performance gains in using hybrid architecture
compared to conventional methods such as Markov switching ARIMA and LSTM.
Related papers
- Recursive Learning of Asymptotic Variational Objectives [49.69399307452126]
General state-space models (SSMs) are widely used in statistical machine learning and are among the most classical generative models for sequential time-series data.
Online sequential IWAE (OSIWAE) allows for online learning of both model parameters and a Markovian recognition model for inferring latent states.
This approach is more theoretically well-founded than recently proposed online variational SMC methods.
arXiv Detail & Related papers (2024-11-04T16:12:37Z) - Synthetic location trajectory generation using categorical diffusion
models [50.809683239937584]
Diffusion models (DPMs) have rapidly evolved to be one of the predominant generative models for the simulation of synthetic data.
We propose using DPMs for the generation of synthetic individual location trajectories (ILTs) which are sequences of variables representing physical locations visited by individuals.
arXiv Detail & Related papers (2024-02-19T15:57:39Z) - Latent State Models of Training Dynamics [51.88132043461152]
We train models with different random seeds and compute a variety of metrics throughout training.
We then fit a hidden Markov model (HMM) over the resulting sequences of metrics.
We use the HMM representation to study phase transitions and identify latent "detour" states that slow down convergence.
arXiv Detail & Related papers (2023-08-18T13:20:08Z) - Switching Autoregressive Low-rank Tensor Models [12.461139675114818]
We show how to switch autoregressive low-rank tensor (SALT) models.
SALT parameterizes the tensor of an ARHMM with a low-rank factorization to control the number of parameters.
We prove theoretical and discuss practical connections between SALT, linear dynamical systems, and SLDSs.
arXiv Detail & Related papers (2023-06-05T22:25:28Z) - Formal Controller Synthesis for Markov Jump Linear Systems with
Uncertain Dynamics [64.72260320446158]
We propose a method for synthesising controllers for Markov jump linear systems.
Our method is based on a finite-state abstraction that captures both the discrete (mode-jumping) and continuous (stochastic linear) behaviour of the MJLS.
We apply our method to multiple realistic benchmark problems, in particular, a temperature control and an aerial vehicle delivery problem.
arXiv Detail & Related papers (2022-12-01T17:36:30Z) - Realization of the Trajectory Propagation in the MM-SQC Dynamics by
Using Machine Learning [4.629634111796585]
We apply the supervised machine learning (ML) approach to realize the trajectory-based nonadiabatic dynamics.
The proposed idea is proven to be reliable and accurate in the simulations of the dynamics of several site-exciton electron-phonon coupling models.
arXiv Detail & Related papers (2022-07-11T01:23:36Z) - Understanding Dynamic Spatio-Temporal Contexts in Long Short-Term Memory
for Road Traffic Speed Prediction [11.92436948211501]
We propose a dynamically localised long short-term memory (LSTM) model that involves both spatial and temporal dependence between roads.
The LSTM model can deal with sequential data with long dependency as well as complex non-linear features.
Empirical results indicated superior prediction performances of the proposed model compared to two different baseline methods.
arXiv Detail & Related papers (2021-12-04T19:33:05Z) - Equivalence of Segmental and Neural Transducer Modeling: A Proof of
Concept [56.46135010588918]
We prove that the widely used class of RNN-Transducer models and segmental models (direct HMM) are equivalent.
It is shown that blank probabilities translate into segment length probabilities and vice versa.
arXiv Detail & Related papers (2021-04-13T11:20:48Z) - Dynamic Gaussian Mixture based Deep Generative Model For Robust
Forecasting on Sparse Multivariate Time Series [43.86737761236125]
We propose a novel generative model, which tracks the transition of latent clusters, instead of isolated feature representations.
It is characterized by a newly designed dynamic Gaussian mixture distribution, which captures the dynamics of clustering structures.
A structured inference network is also designed for enabling inductive analysis.
arXiv Detail & Related papers (2021-03-03T04:10:07Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Markovian RNN: An Adaptive Time Series Prediction Network with HMM-based
Switching for Nonstationary Environments [11.716677452529114]
We introduce a novel recurrent neural network (RNN) architecture, which adaptively switches between internal regimes in a Markovian way to model the nonstationary nature of the given data.
Our model, Markovian RNN employs a hidden Markov model (HMM) for regime transitions, where each regime controls hidden state transitions of the recurrent cell independently.
We demonstrate the significant performance gains compared to vanilla RNN and conventional methods such as Markov Switching ARIMA.
arXiv Detail & Related papers (2020-06-17T19:38:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.