Quantifying Distribution Shift in Traffic Signal Control with Histogram-Based GEH Distance
- URL: http://arxiv.org/abs/2511.13785v1
- Date: Sun, 16 Nov 2025 11:05:02 GMT
- Title: Quantifying Distribution Shift in Traffic Signal Control with Histogram-Based GEH Distance
- Authors: Federico Taschin, Ozan K. Tonguz,
- Abstract summary: This paper introduces a principled approach to quantify distribution shift by representing traffic scenarios as demand histograms.<n>We validate the approach on 20 simulated scenarios using both a NEMA actuated controller and a reinforcement learning controller.<n>Overall, this method can predict performance degradation under distribution shift better than previously published techniques.
- Score: 0.6961253535504978
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Traffic signal control algorithms are vulnerable to distribution shift, where performance degrades under traffic conditions that differ from those seen during design or training. This paper introduces a principled approach to quantify distribution shift by representing traffic scenarios as demand histograms and comparing them with a GEH-based distance function. The method is policy-independent, interpretable, and leverages a widely used traffic engineering statistic. We validate the approach on 20 simulated scenarios using both a NEMA actuated controller and a reinforcement learning controller (FRAP++). Results show that larger scenario distances consistently correspond to increased travel time and reduced throughput, with particularly strong explanatory power for learning-based control. Overall, this method can predict performance degradation under distribution shift better than previously published techniques. These findings highlight the utility of the proposed framework for benchmarking, training regime design, and monitoring in adaptive traffic signal control.
Related papers
- Why Steering Works: Toward a Unified View of Language Model Parameter Dynamics [81.80010043113445]
Local weight fine-tuning, LoRA-based adaptation, and activation-based interventions are studied in isolation.<n>We present a unified view that frames these interventions as dynamic weight updates induced by a control signal.<n>Across methods, we observe a consistent trade-off between preference and utility: stronger control increases preference while predictably reducing utility.
arXiv Detail & Related papers (2026-02-02T17:04:36Z) - Network-Level Vehicle Delay Estimation at Heterogeneous Signalized Intersections [4.534054317956599]
This study introduces a domain adaptation (DA) framework for estimating vehicle delays across diverse intersections.<n>A novel DA model, Gradient Boosting with Balanced Weighting (GBBW), reweights source data based on similarity to the target domain.<n>Performance is evaluated against eight state-of-the-art ML regression models and seven instance-based DA methods.
arXiv Detail & Related papers (2025-10-01T05:19:50Z) - Reinforcement Learning Based Traffic Signal Design to Minimize Queue Lengths [15.439906983758808]
We propose a novel adaptive TSC framework that leverages Reinforcement Learning (RL) to minimize total queue lengths across all signal phases.<n>The proposed algorithm has been implemented using the Simulation of Urban Mobility (SUMO) traffic simulator.<n>The best performing configuration achieves an approximately 29% reduction in average queue lengths compared to the traditional Webster method.
arXiv Detail & Related papers (2025-09-26T01:23:30Z) - Enhancing Traffic Signal Control through Model-based Reinforcement Learning and Policy Reuse [0.9995933996287355]
Multi-agent reinforcement learning (MARL) has shown significant potential in traffic signal control (TSC)<n>Current MARL-based methods often suffer from insufficient generalization due to the fixed traffic patterns and road network conditions used during training.<n>This limitation results in poor adaptability to new traffic scenarios, leading to high retraining costs and complex deployment.<n>We propose two algorithms: PLight and PRLight. PLight employs a model-based reinforcement learning approach, pretraining control policies and environment models using predefined source-domain traffic scenarios. PRLight further enhances adaptability by adaptively selecting pre-trained PLight agents based on the similarity between
arXiv Detail & Related papers (2025-03-11T01:21:13Z) - Improving Traffic Flow Predictions with SGCN-LSTM: A Hybrid Model for Spatial and Temporal Dependencies [55.2480439325792]
This paper introduces the Signal-Enhanced Graph Convolutional Network Long Short Term Memory (SGCN-LSTM) model for predicting traffic speeds across road networks.
Experiments on the PEMS-BAY road network traffic dataset demonstrate the SGCN-LSTM model's effectiveness.
arXiv Detail & Related papers (2024-11-01T00:37:00Z) - Strada-LLM: Graph LLM for traffic prediction [62.2015839597764]
A considerable challenge in traffic prediction lies in handling the diverse data distributions caused by vastly different traffic conditions.<n>We propose a graph-aware LLM for traffic prediction that considers proximal traffic information.<n>We adopt a lightweight approach for efficient domain adaptation when facing new data distributions in few-shot fashion.
arXiv Detail & Related papers (2024-10-28T09:19:29Z) - A Holistic Framework Towards Vision-based Traffic Signal Control with
Microscopic Simulation [53.39174966020085]
Traffic signal control (TSC) is crucial for reducing traffic congestion that leads to smoother traffic flow, reduced idling time, and mitigated CO2 emissions.
In this study, we explore the computer vision approach for TSC that modulates on-road traffic flows through visual observation.
We introduce a holistic traffic simulation framework called TrafficDojo towards vision-based TSC and its benchmarking.
arXiv Detail & Related papers (2024-03-11T16:42:29Z) - Lyapunov Function Consistent Adaptive Network Signal Control with Back
Pressure and Reinforcement Learning [9.797994846439527]
This study introduces a unified framework using Lyapunov control theory, defining specific Lyapunov functions respectively.
Building on insights from Lyapunov theory, this study designs a reward function for the Reinforcement Learning (RL)-based network signal control.
The proposed algorithm is compared with several traditional and RL-based methods under pure passenger car flow and heterogenous traffic flow including freight.
arXiv Detail & Related papers (2022-10-06T00:22:02Z) - Time-to-Green predictions for fully-actuated signal control systems with
supervised learning [56.66331540599836]
This paper proposes a time series prediction framework using aggregated traffic signal and loop detector data.
We utilize state-of-the-art machine learning models to predict future signal phases' duration.
Results based on an empirical data set from a fully-actuated signal control system in Zurich, Switzerland, show that machine learning models outperform conventional prediction methods.
arXiv Detail & Related papers (2022-08-24T07:50:43Z) - A Deep Reinforcement Learning Approach for Traffic Signal Control
Optimization [14.455497228170646]
Inefficient traffic signal control methods may cause numerous problems, such as traffic congestion and waste of energy.
This paper first proposes a multi-agent deep deterministic policy gradient (MADDPG) method by extending the actor-critic policy gradient algorithms.
arXiv Detail & Related papers (2021-07-13T14:11:04Z) - Federated Learning on the Road: Autonomous Controller Design for
Connected and Autonomous Vehicles [109.71532364079711]
A new federated learning (FL) framework is proposed for designing the autonomous controller of connected and autonomous vehicles (CAVs)
A novel dynamic federated proximal (DFP) algorithm is proposed that accounts for the mobility of CAVs, the wireless fading channels, and the unbalanced and nonindependent and identically distributed data across CAVs.
A rigorous convergence analysis is performed for the proposed algorithm to identify how fast the CAVs converge to using the optimal controller.
arXiv Detail & Related papers (2021-02-05T19:57:47Z) - MetaVIM: Meta Variationally Intrinsic Motivated Reinforcement Learning for Decentralized Traffic Signal Control [54.162449208797334]
Traffic signal control aims to coordinate traffic signals across intersections to improve the traffic efficiency of a district or a city.
Deep reinforcement learning (RL) has been applied to traffic signal control recently and demonstrated promising performance where each traffic signal is regarded as an agent.
We propose a novel Meta Variationally Intrinsic Motivated (MetaVIM) RL method to learn the decentralized policy for each intersection that considers neighbor information in a latent way.
arXiv Detail & Related papers (2021-01-04T03:06:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.