Multi-Scale Diffusion Transformer for Jointly Simulating User Mobility and Mobile Traffic Pattern
- URL: http://arxiv.org/abs/2510.10158v1
- Date: Sat, 11 Oct 2025 10:45:39 GMT
- Title: Multi-Scale Diffusion Transformer for Jointly Simulating User Mobility and Mobile Traffic Pattern
- Authors: Ziyi Liu, Qingyue Long, Zhiwen Xue, Huandong Wang, Yong Li,
- Abstract summary: MSTDiff is a multi-scale Diffusion Transformer for joint simulation of mobile traffic and user trajectories.<n>It surpasses state-of-the-art baselines in traffic and trajectory generation tasks.
- Score: 20.211463364436202
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: User mobility trajectory and mobile traffic data are essential for a wide spectrum of applications including urban planning, network optimization, and emergency management. However, large-scale and fine-grained mobility data remains difficult to obtain due to privacy concerns and collection costs, making it essential to simulate realistic mobility and traffic patterns. User trajectories and mobile traffic are fundamentally coupled, reflecting both physical mobility and cyber behavior in urban environments. Despite this strong interdependence, existing studies often model them separately, limiting the ability to capture cross-modal dynamics. Therefore, a unified framework is crucial. In this paper, we propose MSTDiff, a Multi-Scale Diffusion Transformer for joint simulation of mobile traffic and user trajectories. First, MSTDiff applies discrete wavelet transforms for multi-resolution traffic decomposition. Second, it uses a hybrid denoising network to process continuous traffic volumes and discrete location sequences. A transition mechanism based on urban knowledge graph embedding similarity is designed to guide semantically informed trajectory generation. Finally, a multi-scale Transformer with cross-attention captures dependencies between trajectories and traffic. Experiments show that MSTDiff surpasses state-of-the-art baselines in traffic and trajectory generation tasks, reducing Jensen-Shannon divergence (JSD) across key statistical metrics by up to 17.38% for traffic generation, and by an average of 39.53% for trajectory generation. The source code is available at: https://github.com/tsinghua-fib-lab/MSTDiff .
Related papers
- Spatiotemporal Decision Transformer for Traffic Coordination [1.2099551931618155]
MADT (Multi-Agent Decision Transformer) is a novel approach that reformulates multi-agent traffic signal control as a sequence modeling problem.<n>Our approach enables offline learning from historical traffic data, with architecture design that facilitates potential online fine-tuning.
arXiv Detail & Related papers (2026-02-02T23:19:13Z) - Denoising Refinement Diffusion Models for Simultaneous Generation of Multi-scale Mobile Network Traffic [7.300362420979762]
We propose ZoomDiff, a diffusion-based model for multi-scale mobile traffic generation.<n>DRDM employs a multi-stage noise-adding and denoising mechanism, enabling different stages to generate traffic attemporal resolutions.<n>Experiments on real-world mobile traffic datasets show that ZoomDiff at least an 18.4% improvement over state-of-the-art baselines in multi-scale traffic generation tasks.
arXiv Detail & Related papers (2025-10-30T03:42:56Z) - Rethinking Traffic Flow Forecasting: From Transition to Generatation [0.0]
We propose an Effective Multi-Branch Similarity Transformer for Traffic Flow Prediction, namely EMBSFormer.<n>We find that the factors affecting traffic flow include node-level traffic generation and graph-level traffic transition, which describe the multi-periodicity and interaction pattern of nodes, respectively.<n>For traffic transition, we employ a temporal and spatial self-attention mechanism to maintain global node interactions, and use GNN and time conv to model local node interactions, respectively.
arXiv Detail & Related papers (2025-04-19T09:52:39Z) - Improving Traffic Flow Predictions with SGCN-LSTM: A Hybrid Model for Spatial and Temporal Dependencies [55.2480439325792]
This paper introduces the Signal-Enhanced Graph Convolutional Network Long Short Term Memory (SGCN-LSTM) model for predicting traffic speeds across road networks.
Experiments on the PEMS-BAY road network traffic dataset demonstrate the SGCN-LSTM model's effectiveness.
arXiv Detail & Related papers (2024-11-01T00:37:00Z) - Deep Learning-driven Mobile Traffic Measurement Collection and Analysis [0.43512163406552007]
In this thesis, we harness the powerful hierarchical feature learning abilities of Deep Learning (DL) techniques in both spatial and temporal domains.
We develop solutions for precise city-scale mobile traffic analysis and forecasting.
arXiv Detail & Related papers (2024-10-14T06:53:45Z) - Rethinking Urban Mobility Prediction: A Super-Multivariate Time Series
Forecasting Approach [71.67506068703314]
Long-term urban mobility predictions play a crucial role in the effective management of urban facilities and services.
Traditionally, urban mobility data has been structured as videos, treating longitude and latitude as fundamental pixels.
In our research, we introduce a fresh perspective on urban mobility prediction.
Instead of oversimplifying urban mobility data as traditional video data, we regard it as a complex time series.
arXiv Detail & Related papers (2023-12-04T07:39:05Z) - Communication-Efficient Framework for Distributed Image Semantic
Wireless Transmission [68.69108124451263]
Federated learning-based semantic communication (FLSC) framework for multi-task distributed image transmission with IoT devices.
Each link is composed of a hierarchical vision transformer (HVT)-based extractor and a task-adaptive translator.
Channel state information-based multiple-input multiple-output transmission module designed to combat channel fading and noise.
arXiv Detail & Related papers (2023-08-07T16:32:14Z) - PDFormer: Propagation Delay-Aware Dynamic Long-Range Transformer for
Traffic Flow Prediction [78.05103666987655]
spatial-temporal Graph Neural Network (GNN) models have emerged as one of the most promising methods to solve this problem.
We propose a novel propagation delay-aware dynamic long-range transFormer, namely PDFormer, for accurate traffic flow prediction.
Our method can not only achieve state-of-the-art performance but also exhibit competitive computational efficiency.
arXiv Detail & Related papers (2023-01-19T08:42:40Z) - Decoupled Dynamic Spatial-Temporal Graph Neural Network for Traffic
Forecasting [27.82230529014677]
The ability to forecast the state of traffic in a road network is an important functionality and a challenging task.
Recent proposals on spatial-temporal graph neural networks have achieved great progress at modeling complex spatial-temporal correlations in traffic data.
We propose a novel Decoupled Spatial-Temporal Framework (DSTF) that separates the diffusion and inherent traffic information in a data-driven manner.
arXiv Detail & Related papers (2022-06-18T04:14:38Z) - Short-term passenger flow prediction for multi-traffic modes: A residual
network and Transformer based multi-task learning method [21.13073816634534]
Res-Transformer is a learning model for short-term passenger flow prediction of multi-traffic modes.
Model is evaluated on two large-scale real-world datasets from Beijing, China.
This paper can give critical insights into the short-tern passenger flow prediction for multi-traffic modes.
arXiv Detail & Related papers (2022-02-27T01:09:19Z) - Road Network Guided Fine-Grained Urban Traffic Flow Inference [108.64631590347352]
Accurate inference of fine-grained traffic flow from coarse-grained one is an emerging yet crucial problem.
We propose a novel Road-Aware Traffic Flow Magnifier (RATFM) that exploits the prior knowledge of road networks.
Our method can generate high-quality fine-grained traffic flow maps.
arXiv Detail & Related papers (2021-09-29T07:51:49Z) - Learning dynamic and hierarchical traffic spatiotemporal features with
Transformer [4.506591024152763]
This paper proposes a novel model, Traffic Transformer, for spatial-temporal graph modeling and long-term traffic forecasting.
Transformer is the most popular framework in Natural Language Processing (NLP)
analyzing the attention weight matrixes can find the influential part of road networks, allowing us to learn the traffic networks better.
arXiv Detail & Related papers (2021-04-12T02:29:58Z) - Multi-intersection Traffic Optimisation: A Benchmark Dataset and a
Strong Baseline [85.9210953301628]
Control of traffic signals is fundamental and critical to alleviate traffic congestion in urban areas.
Because of the high complexity of modelling the problem, experimental settings of current works are often inconsistent.
We propose a novel and strong baseline model based on deep reinforcement learning with the encoder-decoder structure.
arXiv Detail & Related papers (2021-01-24T03:55:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.