Dynamic Campus Origin-Destination Mobility Prediction using Graph Convolutional Neural Network on WiFi Logs
- URL: http://arxiv.org/abs/2507.05507v1
- Date: Mon, 07 Jul 2025 22:04:43 GMT
- Title: Dynamic Campus Origin-Destination Mobility Prediction using Graph Convolutional Neural Network on WiFi Logs
- Authors: Godwin Badu-Marfo, Bilal Farooq,
- Abstract summary: We present an integrated graph-based neural networks architecture for predicting campus buildings occupancy and inter-buildings movement.<n>We learn traffic flow patterns from Wi-Fi logs combined with the usage schedules within the buildings.<n>The results of the experiments show that the integrated GTM models significantly outperform traditional pedestrian flow estimators.
- Score: 5.759608579971381
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present an integrated graph-based neural networks architecture for predicting campus buildings occupancy and inter-buildings movement at dynamic temporal resolution that learns traffic flow patterns from Wi-Fi logs combined with the usage schedules within the buildings. The relative traffic flows are directly estimated from the WiFi data without assuming the occupant behaviour or preferences while maintaining individual privacy. We formulate the problem as a data-driven graph structure represented by a set of nodes (representing buildings), connected through a route of edges or links using a novel Graph Convolution plus LSTM Neural Network (GCLSTM) which has shown remarkable success in modelling complex patterns. We describe the formulation, model estimation, interpretability and examine the relative performance of our proposed model. We also present an illustrative architecture of the models and apply on real-world WiFi logs collected at the Toronto Metropolitan University campus. The results of the experiments show that the integrated GCLSTM models significantly outperform traditional pedestrian flow estimators like the Multi Layer Perceptron (MLP) and Linear Regression.
Related papers
- Joint Graph Convolution and Sequential Modeling for Scalable Network Traffic Estimation [11.751952500567388]
This study focuses on the challenge of predicting network traffic within complex topological environments.<n>It introduces a Graph Contemporalal Networks (GCN) with Gated Recurrent Units (GRU) modeling approach.<n>The effectiveness of the proposed model is validated through comprehensive experiments on the real-world Abilene network traffic dataset.
arXiv Detail & Related papers (2025-05-12T15:38:19Z) - Generalized Factor Neural Network Model for High-dimensional Regression [50.554377879576066]
We tackle the challenges of modeling high-dimensional data sets with latent low-dimensional structures hidden within complex, non-linear, and noisy relationships.<n>Our approach enables a seamless integration of concepts from non-parametric regression, factor models, and neural networks for high-dimensional regression.
arXiv Detail & Related papers (2025-02-16T23:13:55Z) - UniFlow: A Foundation Model for Unified Urban Spatio-Temporal Flow Prediction [23.79904267354509]
Urban-temporal flow prediction is crucial for optimizing city infrastructure and managing traffic emergency responses.<n>Traditional approaches have relied on separate models tailored to either grid-based data, or graph-based data.<n>In this paper we propose a model for general urban flow prediction that unifies both grid-based graphbased data.
arXiv Detail & Related papers (2024-11-20T01:54:52Z) - Improving Traffic Flow Predictions with SGCN-LSTM: A Hybrid Model for Spatial and Temporal Dependencies [55.2480439325792]
This paper introduces the Signal-Enhanced Graph Convolutional Network Long Short Term Memory (SGCN-LSTM) model for predicting traffic speeds across road networks.
Experiments on the PEMS-BAY road network traffic dataset demonstrate the SGCN-LSTM model's effectiveness.
arXiv Detail & Related papers (2024-11-01T00:37:00Z) - Contextualizing MLP-Mixers Spatiotemporally for Urban Data Forecast at Scale [54.15522908057831]
We propose an adapted version of the computationally-Mixer for STTD forecast at scale.
Our results surprisingly show that this simple-yeteffective solution can rival SOTA baselines when tested on several traffic benchmarks.
Our findings contribute to the exploration of simple-yet-effective models for real-world STTD forecasting.
arXiv Detail & Related papers (2023-07-04T05:19:19Z) - Simpler is better: Multilevel Abstraction with Graph Convolutional
Recurrent Neural Network Cells for Traffic Prediction [6.968068088508505]
We present a new sequence-to-sequence architecture for graph neural networks (GNNs)
We also present a new benchmark benchmark dataset of street-level segment data in Montreal, Canada.
Our model improves performance by more than 7% for one-hour prediction compared to the baseline methods.
arXiv Detail & Related papers (2022-09-08T14:56:29Z) - A spatial-temporal short-term traffic flow prediction model based on
dynamical-learning graph convolution mechanism [0.0]
Short-term traffic flow prediction is a vital branch of the Intelligent Traffic System (ITS) and plays an important role in traffic management.
Graph convolution network (GCN) is widely used in traffic prediction models to better deal with the graphical structure data of road networks.
To deal with this drawback, this paper proposes a novel location graph convolutional network (Location-GCN)
arXiv Detail & Related papers (2022-05-10T09:19:12Z) - Spatio-Temporal Latent Graph Structure Learning for Traffic Forecasting [6.428566223253948]
We propose a new traffic forecasting framework--S-Temporal Latent Graph Structure Learning networks (ST-LGSL)
The model employs a graph based on Multilayer perceptron and K-Nearest Neighbor, which learns the latent graph topological information from the entire data.
With the dependencies-kNN based on ground-truth adjacency matrix and similarity metric in kNN, ST-LGSL aggregates the top focusing on geography and node similarity.
arXiv Detail & Related papers (2022-02-25T10:02:49Z) - BScNets: Block Simplicial Complex Neural Networks [79.81654213581977]
Simplicial neural networks (SNN) have recently emerged as the newest direction in graph learning.
We present Block Simplicial Complex Neural Networks (BScNets) model for link prediction.
BScNets outperforms state-of-the-art models by a significant margin while maintaining low costs.
arXiv Detail & Related papers (2021-12-13T17:35:54Z) - TCL: Transformer-based Dynamic Graph Modelling via Contrastive Learning [87.38675639186405]
We propose a novel graph neural network approach, called TCL, which deals with the dynamically-evolving graph in a continuous-time fashion.
To the best of our knowledge, this is the first attempt to apply contrastive learning to representation learning on dynamic graphs.
arXiv Detail & Related papers (2021-05-17T15:33:25Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - A Generative Learning Approach for Spatio-temporal Modeling in Connected
Vehicular Network [55.852401381113786]
This paper proposes LaMI (Latency Model Inpainting), a novel framework to generate a comprehensive-temporal quality framework for wireless access latency of connected vehicles.
LaMI adopts the idea from image inpainting and synthesizing and can reconstruct the missing latency samples by a two-step procedure.
In particular, it first discovers the spatial correlation between samples collected in various regions using a patching-based approach and then feeds the original and highly correlated samples into a Varienational Autocoder (VAE)
arXiv Detail & Related papers (2020-03-16T03:43:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.