Independent Reinforcement Learning for Weakly Cooperative Multiagent
Traffic Control Problem
- URL: http://arxiv.org/abs/2104.10917v1
- Date: Thu, 22 Apr 2021 07:55:46 GMT
- Title: Independent Reinforcement Learning for Weakly Cooperative Multiagent
Traffic Control Problem
- Authors: Chengwei Zhang and Shan Jin and Wanli Xue and Xiaofei Xie and
Shengyong Chen and Rong Chen
- Abstract summary: We use independent reinforcement learning (IRL) to solve a complex traffic cooperative control problem in this study.
To this, we model the traffic control problem as a partially observable weak cooperative traffic model (PO-WCTM) to optimize the overall traffic situation of a group of intersections.
Experimental results show that CIL-DDQN outperforms other methods in almost all performance indicators of the traffic control problem.
- Score: 22.733542222812158
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The adaptive traffic signal control (ATSC) problem can be modeled as a
multiagent cooperative game among urban intersections, where intersections
cooperate to optimize their common goal. Recently, reinforcement learning (RL)
has achieved marked successes in managing sequential decision making problems,
which motivates us to apply RL in the ASTC problem. Here we use independent
reinforcement learning (IRL) to solve a complex traffic cooperative control
problem in this study. One of the largest challenges of this problem is that
the observation information of intersection is typically partially observable,
which limits the learning performance of IRL algorithms. To this, we model the
traffic control problem as a partially observable weak cooperative traffic
model (PO-WCTM) to optimize the overall traffic situation of a group of
intersections. Different from a traditional IRL task that averages the returns
of all agents in fully cooperative games, the learning goal of each
intersection in PO-WCTM is to reduce the cooperative difficulty of learning,
which is also consistent with the traffic environment hypothesis. We also
propose an IRL algorithm called Cooperative Important Lenient Double DQN
(CIL-DDQN), which extends Double DQN (DDQN) algorithm using two mechanisms: the
forgetful experience mechanism and the lenient weight training mechanism. The
former mechanism decreases the importance of experiences stored in the
experience reply buffer, which deals with the problem of experience failure
caused by the strategy change of other agents. The latter mechanism increases
the weight experiences with high estimation and `leniently' trains the DDQN
neural network, which improves the probability of the selection of cooperative
joint strategies. Experimental results show that CIL-DDQN outperforms other
methods in almost all performance indicators of the traffic control problem.
Related papers
- IntersectionZoo: Eco-driving for Benchmarking Multi-Agent Contextual Reinforcement Learning [4.80862277413422]
We propose IntersectionZoo, a comprehensive benchmark suite for multi-agent reinforcement learning.
By grounding IntersectionZoo in a real-world application, we naturally capture real-world problem characteristics.
IntersectionZoo is built on data-informed simulations of 16,334 signalized intersections from 10 major US cities.
arXiv Detail & Related papers (2024-10-19T21:34:24Z) - Towards Interactive and Learnable Cooperative Driving Automation: a Large Language Model-Driven Decision-Making Framework [79.088116316919]
Connected Autonomous Vehicles (CAVs) have begun to open road testing around the world, but their safety and efficiency performance in complex scenarios is still not satisfactory.
This paper proposes CoDrivingLLM, an interactive and learnable LLM-driven cooperative driving framework.
arXiv Detail & Related papers (2024-09-19T14:36:00Z) - Combat Urban Congestion via Collaboration: Heterogeneous GNN-based MARL
for Coordinated Platooning and Traffic Signal Control [16.762073265205565]
This paper proposes an innovative solution to tackle these challenges based on heterogeneous graph multi-agent reinforcement learning and traffic theories.
Our approach involves: 1) designing platoon and signal control as distinct reinforcement learning agents with their own set of observations, actions, and reward functions to optimize traffic flow; 2) designing coordination by incorporating graph neural networks within multi-agent reinforcement learning to facilitate seamless information exchange among agents on a regional scale.
arXiv Detail & Related papers (2023-10-17T02:46:04Z) - Learning to Sail Dynamic Networks: The MARLIN Reinforcement Learning
Framework for Congestion Control in Tactical Environments [53.08686495706487]
This paper proposes an RL framework that leverages an accurate and parallelizable emulation environment to reenact the conditions of a tactical network.
We evaluate our RL learning framework by training a MARLIN agent in conditions replicating a bottleneck link transition between a Satellite Communication (SATCOM) and an UHF Wide Band (UHF) radio link.
arXiv Detail & Related papers (2023-06-27T16:15:15Z) - A Novel Multi-Agent Deep RL Approach for Traffic Signal Control [13.927155702352131]
We propose a Friend-Deep Q-network (Friend-DQN) approach for multiple traffic signal control in urban networks.
In particular, the cooperation between multiple agents can reduce the state-action space and thus speed up the convergence.
arXiv Detail & Related papers (2023-06-05T08:20:37Z) - MARLIN: Soft Actor-Critic based Reinforcement Learning for Congestion
Control in Real Networks [63.24965775030673]
We propose a novel Reinforcement Learning (RL) approach to design generic Congestion Control (CC) algorithms.
Our solution, MARLIN, uses the Soft Actor-Critic algorithm to maximize both entropy and return.
We trained MARLIN on a real network with varying background traffic patterns to overcome the sim-to-real mismatch.
arXiv Detail & Related papers (2023-02-02T18:27:20Z) - Large-Scale Traffic Signal Control by a Nash Deep Q-network Approach [7.23135508361981]
We introduce an off-policy nash deep Q-Network (OPNDQN) algorithm, which mitigates the weakness of both fully centralized and MARL approaches.
One of main advantages of OPNDQN is to mitigate the non-stationarity of multi-agent Markov process.
We show the dominant superiority of OPNDQN over several existing MARL approaches in terms of average queue length, episode training reward and average waiting time.
arXiv Detail & Related papers (2023-01-02T12:58:51Z) - End-to-End Intersection Handling using Multi-Agent Deep Reinforcement
Learning [63.56464608571663]
Navigating through intersections is one of the main challenging tasks for an autonomous vehicle.
In this work, we focus on the implementation of a system able to navigate through intersections where only traffic signs are provided.
We propose a multi-agent system using a continuous, model-free Deep Reinforcement Learning algorithm used to train a neural network for predicting both the acceleration and the steering angle at each time step.
arXiv Detail & Related papers (2021-04-28T07:54:40Z) - Multi-intersection Traffic Optimisation: A Benchmark Dataset and a
Strong Baseline [85.9210953301628]
Control of traffic signals is fundamental and critical to alleviate traffic congestion in urban areas.
Because of the high complexity of modelling the problem, experimental settings of current works are often inconsistent.
We propose a novel and strong baseline model based on deep reinforcement learning with the encoder-decoder structure.
arXiv Detail & Related papers (2021-01-24T03:55:39Z) - Optimizing Mixed Autonomy Traffic Flow With Decentralized Autonomous
Vehicles and Multi-Agent RL [63.52264764099532]
We study the ability of autonomous vehicles to improve the throughput of a bottleneck using a fully decentralized control scheme in a mixed autonomy setting.
We apply multi-agent reinforcement algorithms to this problem and demonstrate that significant improvements in bottleneck throughput, from 20% at a 5% penetration rate to 33% at a 40% penetration rate, can be achieved.
arXiv Detail & Related papers (2020-10-30T22:06:05Z) - Area-wide traffic signal control based on a deep graph Q-Network (DGQN)
trained in an asynchronous manner [3.655021726150368]
Reinforcement learning (RL) algorithms have been widely applied in traffic signal studies.
There are, however, several problems in jointly controlling traffic lights for a large transportation network.
arXiv Detail & Related papers (2020-08-05T06:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.