Decentralized Deep Reinforcement Learning for Network Level Traffic
Signal Control
- URL: http://arxiv.org/abs/2007.03433v2
- Date: Fri, 17 Jul 2020 23:15:45 GMT
- Title: Decentralized Deep Reinforcement Learning for Network Level Traffic
Signal Control
- Authors: Jin Guo
- Abstract summary: I propose a family of fully decentralized deep multi-agent reinforcement learning (MARL) algorithms to achieve high, real-time performance in traffic signal control.
Each intersection is modeled as an agent that plays a Markovian Game against the other intersection nodes.
Experiment results show that S2R2L has a quicker convergence rate and better convergent performance than IDQL and S2RL in the training process.
- Score: 0.8838408191955874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this thesis, I propose a family of fully decentralized deep multi-agent
reinforcement learning (MARL) algorithms to achieve high, real-time performance
in network-level traffic signal control. In this approach, each intersection is
modeled as an agent that plays a Markovian Game against the other intersection
nodes in a traffic signal network modeled as an undirected graph, to approach
the optimal reduction in delay. Following Partially Observable Markov Decision
Processes (POMDPs), there are 3 levels of communication schemes between
adjacent learning agents: independent deep Q-leaning (IDQL), shared states
reinforcement learning (S2RL) and a shared states & rewards version of
S2RL--S2R2L. In these 3 variants of decentralized MARL schemes, individual
agent trains its local deep Q network (DQN) separately, enhanced by
convergence-guaranteed techniques like double DQN, prioritized experience
replay, multi-step bootstrapping, etc. To test the performance of the proposed
three MARL algorithms, a SUMO-based simulation platform is developed to mimic
the traffic evolution of the real world. Fed with random traffic demand between
permitted OD pairs, a 4x4 Manhattan-style grid network is set up as the
testbed, two different vehicle arrival rates are generated for model training
and testing. The experiment results show that S2R2L has a quicker convergence
rate and better convergent performance than IDQL and S2RL in the training
process. Moreover, three MARL schemes all reveal exceptional generalization
abilities. Their testing results surpass the benchmark Max Pressure (MP)
algorithm, under the criteria of average vehicle delay, network-level queue
length and fuel consumption rate. Notably, S2R2L has the best testing
performance of reducing 34.55% traffic delay and dissipating 10.91% queue
length compared with MP.
Related papers
- Applying Neural Monte Carlo Tree Search to Unsignalized Multi-intersection Scheduling for Autonomous Vehicles [7.32653612106583]
We introduce a transformation model that maps sequences of potentially conflicting road-space reservation requests from platoons of vehicles into a series of board-game-like problems.
We use NMCTS to search for solutions representing optimal road-space allocation schedules in the context of past allocations.
We show that the proposed method maintained free-flow in light traffic when all intersections are under control of PNMCTS and outperformed state-of-the-art RL-based traffic-light controllers in average travel time by 74.5% and total throughput by 16% in heavy traffic.
arXiv Detail & Related papers (2024-10-24T14:37:55Z) - FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - SPARQ: Efficient Entanglement Distribution and Routing in Space-Air-Ground Quantum Networks [50.91365514137301]
Space-air-ground quantum (SPARQ) network is developed as a means for providing a seamless on-demand entanglement distribution.
Deep reinforcement learning framework is proposed and trained using deep Q-network (DQN) on multiple graphs of SPARQ.
Third-party entanglement distribution policy is proposed to establish entanglement between communication parties.
arXiv Detail & Related papers (2024-09-19T16:31:37Z) - Active search and coverage using point-cloud reinforcement learning [50.741409008225766]
This paper presents an end-to-end deep reinforcement learning solution for target search and coverage.
We show that deep hierarchical feature learning works for RL and that by using farthest point sampling (FPS) we can reduce the amount of points.
We also show that multi-head attention for point-clouds helps to learn the agent faster but converges to the same outcome.
arXiv Detail & Related papers (2023-12-18T18:16:30Z) - Large-Scale Traffic Signal Control by a Nash Deep Q-network Approach [7.23135508361981]
We introduce an off-policy nash deep Q-Network (OPNDQN) algorithm, which mitigates the weakness of both fully centralized and MARL approaches.
One of main advantages of OPNDQN is to mitigate the non-stationarity of multi-agent Markov process.
We show the dominant superiority of OPNDQN over several existing MARL approaches in terms of average queue length, episode training reward and average waiting time.
arXiv Detail & Related papers (2023-01-02T12:58:51Z) - Optimizing Mixed Autonomy Traffic Flow With Decentralized Autonomous
Vehicles and Multi-Agent RL [63.52264764099532]
We study the ability of autonomous vehicles to improve the throughput of a bottleneck using a fully decentralized control scheme in a mixed autonomy setting.
We apply multi-agent reinforcement algorithms to this problem and demonstrate that significant improvements in bottleneck throughput, from 20% at a 5% penetration rate to 33% at a 40% penetration rate, can be achieved.
arXiv Detail & Related papers (2020-10-30T22:06:05Z) - Multi-Agent Reinforcement Learning in NOMA-aided UAV Networks for
Cellular Offloading [59.32570888309133]
A novel framework is proposed for cellular offloading with the aid of multiple unmanned aerial vehicles (UAVs)
Non-orthogonal multiple access (NOMA) technique is employed at each UAV to further improve the spectrum efficiency of the wireless network.
A mutual deep Q-network (MDQN) algorithm is proposed to jointly determine the optimal 3D trajectory and power allocation of UAVs.
arXiv Detail & Related papers (2020-10-18T20:22:05Z) - Area-wide traffic signal control based on a deep graph Q-Network (DGQN)
trained in an asynchronous manner [3.655021726150368]
Reinforcement learning (RL) algorithms have been widely applied in traffic signal studies.
There are, however, several problems in jointly controlling traffic lights for a large transportation network.
arXiv Detail & Related papers (2020-08-05T06:13:58Z) - Reinforcement Learning Based Vehicle-cell Association Algorithm for
Highly Mobile Millimeter Wave Communication [53.47785498477648]
This paper investigates the problem of vehicle-cell association in millimeter wave (mmWave) communication networks.
We first formulate the user state (VU) problem as a discrete non-vehicle association optimization problem.
The proposed solution achieves up to 15% gains in terms sum of user complexity and 20% reduction in VUE compared to several baseline designs.
arXiv Detail & Related papers (2020-01-22T08:51:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.