Model-based graph reinforcement learning for inductive traffic signal
control
- URL: http://arxiv.org/abs/2208.00659v1
- Date: Mon, 1 Aug 2022 07:43:38 GMT
- Title: Model-based graph reinforcement learning for inductive traffic signal
control
- Authors: Fran\c{c}ois-Xavier Devailly, Denis Larocque, Laurent Charlin
- Abstract summary: Most reinforcement learning methods for adaptive-traffic-signal-control require training from scratch to be applied on any new intersection.
Recent approaches enable learning policies that generalize for unseen road-network topologies and traffic distributions.
We introduce a new model-based method, MuJAM, which on top of enabling explicit coordination at scale for the first time.
- Score: 4.273991039651846
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most reinforcement learning methods for adaptive-traffic-signal-control
require training from scratch to be applied on any new intersection or after
any modification to the road network, traffic distribution, or behavioral
constraints experienced during training. Considering 1) the massive amount of
experience required to train such methods, and 2) that experience must be
gathered by interacting in an exploratory fashion with real road-network-users,
such a lack of transferability limits experimentation and applicability. Recent
approaches enable learning policies that generalize for unseen road-network
topologies and traffic distributions, partially tackling this challenge.
However, the literature remains divided between the learning of cyclic (the
evolution of connectivity at an intersection must respect a cycle) and acyclic
(less constrained) policies, and these transferable methods 1) are only
compatible with cyclic constraints and 2) do not enable coordination. We
introduce a new model-based method, MuJAM, which, on top of enabling explicit
coordination at scale for the first time, pushes generalization further by
allowing a generalization to the controllers' constraints. In a zero-shot
transfer setting involving both road networks and traffic settings never
experienced during training, and in a larger transfer experiment involving the
control of 3,971 traffic signal controllers in Manhattan, we show that MuJAM,
using both cyclic and acyclic constraints, outperforms domain-specific
baselines as well as another transferable approach.
Related papers
- A Holistic Framework Towards Vision-based Traffic Signal Control with
Microscopic Simulation [53.39174966020085]
Traffic signal control (TSC) is crucial for reducing traffic congestion that leads to smoother traffic flow, reduced idling time, and mitigated CO2 emissions.
In this study, we explore the computer vision approach for TSC that modulates on-road traffic flows through visual observation.
We introduce a holistic traffic simulation framework called TrafficDojo towards vision-based TSC and its benchmarking.
arXiv Detail & Related papers (2024-03-11T16:42:29Z) - DenseLight: Efficient Control for Large-scale Traffic Signals with Dense
Feedback [109.84667902348498]
Traffic Signal Control (TSC) aims to reduce the average travel time of vehicles in a road network.
Most prior TSC methods leverage deep reinforcement learning to search for a control policy.
We propose DenseLight, a novel RL-based TSC method that employs an unbiased reward function to provide dense feedback on policy effectiveness.
arXiv Detail & Related papers (2023-06-13T05:58:57Z) - Improving the generalizability and robustness of large-scale traffic
signal control [3.8028221877086814]
We study the robustness of deep reinforcement-learning (RL) approaches to control traffic signals.
We show that recent methods remain brittle in the face of missing data.
We propose using a combination of distributional and vanilla reinforcement learning through a policy ensemble.
arXiv Detail & Related papers (2023-06-02T21:30:44Z) - Reinforcement Learning Approaches for Traffic Signal Control under
Missing Data [5.896742981602458]
In real-world urban scenarios, missing observation of traffic states may frequently occur due to the lack of sensors.
We propose two solutions: the first one imputes the traffic states to enable adaptive control, and the second one imputes both states and rewards to enable adaptive control and the training of RL agents.
arXiv Detail & Related papers (2023-04-21T03:26:33Z) - Let Offline RL Flow: Training Conservative Agents in the Latent Space of
Normalizing Flows [58.762959061522736]
offline reinforcement learning aims to train a policy on a pre-recorded and fixed dataset without any additional environment interactions.
We build upon recent works on learning policies in latent action spaces and use a special form of Normalizing Flows for constructing a generative model.
We evaluate our method on various locomotion and navigation tasks, demonstrating that our approach outperforms recently proposed algorithms.
arXiv Detail & Related papers (2022-11-20T21:57:10Z) - End-to-End Intersection Handling using Multi-Agent Deep Reinforcement
Learning [63.56464608571663]
Navigating through intersections is one of the main challenging tasks for an autonomous vehicle.
In this work, we focus on the implementation of a system able to navigate through intersections where only traffic signs are provided.
We propose a multi-agent system using a continuous, model-free Deep Reinforcement Learning algorithm used to train a neural network for predicting both the acceleration and the steering angle at each time step.
arXiv Detail & Related papers (2021-04-28T07:54:40Z) - Multi-intersection Traffic Optimisation: A Benchmark Dataset and a
Strong Baseline [85.9210953301628]
Control of traffic signals is fundamental and critical to alleviate traffic congestion in urban areas.
Because of the high complexity of modelling the problem, experimental settings of current works are often inconsistent.
We propose a novel and strong baseline model based on deep reinforcement learning with the encoder-decoder structure.
arXiv Detail & Related papers (2021-01-24T03:55:39Z) - MetaVIM: Meta Variationally Intrinsic Motivated Reinforcement Learning for Decentralized Traffic Signal Control [54.162449208797334]
Traffic signal control aims to coordinate traffic signals across intersections to improve the traffic efficiency of a district or a city.
Deep reinforcement learning (RL) has been applied to traffic signal control recently and demonstrated promising performance where each traffic signal is regarded as an agent.
We propose a novel Meta Variationally Intrinsic Motivated (MetaVIM) RL method to learn the decentralized policy for each intersection that considers neighbor information in a latent way.
arXiv Detail & Related papers (2021-01-04T03:06:08Z) - IG-RL: Inductive Graph Reinforcement Learning for Massive-Scale Traffic
Signal Control [4.273991039651846]
Scaling adaptive traffic-signal control involves dealing with state and action spaces.
We introduce Inductive Graph Reinforcement Learning (IG-RL) based on graph-convolutional networks.
Our model can generalize to new road networks, traffic distributions, and traffic regimes.
arXiv Detail & Related papers (2020-03-06T17:17:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.