Optimizing Mixed Autonomy Traffic Flow With Decentralized Autonomous
Vehicles and Multi-Agent RL
- URL: http://arxiv.org/abs/2011.00120v1
- Date: Fri, 30 Oct 2020 22:06:05 GMT
- Title: Optimizing Mixed Autonomy Traffic Flow With Decentralized Autonomous
Vehicles and Multi-Agent RL
- Authors: Eugene Vinitsky, Nathan Lichtle, Kanaad Parvate, Alexandre Bayen
- Abstract summary: We study the ability of autonomous vehicles to improve the throughput of a bottleneck using a fully decentralized control scheme in a mixed autonomy setting.
We apply multi-agent reinforcement algorithms to this problem and demonstrate that significant improvements in bottleneck throughput, from 20% at a 5% penetration rate to 33% at a 40% penetration rate, can be achieved.
- Score: 63.52264764099532
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the ability of autonomous vehicles to improve the throughput of a
bottleneck using a fully decentralized control scheme in a mixed autonomy
setting. We consider the problem of improving the throughput of a scaled model
of the San Francisco-Oakland Bay Bridge: a two-stage bottleneck where four
lanes reduce to two and then reduce to one. Although there is extensive work
examining variants of bottleneck control in a centralized setting, there is
less study of the challenging multi-agent setting where the large number of
interacting AVs leads to significant optimization difficulties for
reinforcement learning methods. We apply multi-agent reinforcement algorithms
to this problem and demonstrate that significant improvements in bottleneck
throughput, from 20\% at a 5\% penetration rate to 33\% at a 40\% penetration
rate, can be achieved. We compare our results to a hand-designed feedback
controller and demonstrate that our results sharply outperform the feedback
controller despite extensive tuning. Additionally, we demonstrate that the
RL-based controllers adopt a robust strategy that works across penetration
rates whereas the feedback controllers degrade immediately upon penetration
rate variation. We investigate the feasibility of both action and observation
decentralization and demonstrate that effective strategies are possible using
purely local sensing. Finally, we open-source our code at
https://github.com/eugenevinitsky/decentralized_bottlenecks.
Related papers
- Multi-agent Assessment with QoS Enhancement for HD Map Updates in a Vehicular Network [8.853779271331508]
Reinforcement Learning (RL) algorithms have been used to address the challenging problems in the offloading process of vehicular ad hoc networks (VANET)
Deep Q-learning (DQN) and Actor-critic at the autonomous vehicle (AV) may lead to an increase in the computational load, causing a heavy burden on the computational devices and higher costs.
In this paper, we assess the scalability of an application utilizing a Q-learning single-agent solution in a distributed multi-agent environment.
arXiv Detail & Related papers (2024-07-31T09:17:09Z) - Multi-Agent Soft Actor-Critic with Global Loss for Autonomous Mobility-on-Demand Fleet Control [1.9503475832401784]
We study a sequential decision-making problem for a profit-maximizing operator of an Autonomous Mobility-on-Demand system.
We employ a multi-agent Soft Actor-Critic algorithm combined with weighted bipartite matching.
We show that our approach outperforms state-of-the-art benchmarks by up to 12.9% for dispatching and up to 38.9% with integrated rebalancing.
arXiv Detail & Related papers (2024-04-10T13:49:20Z) - Mastering the Unsupervised Reinforcement Learning Benchmark from Pixels [112.63440666617494]
Reinforcement learning algorithms can succeed but require large amounts of interactions between the agent and the environment.
We propose a new method to solve it, using unsupervised model-based RL, for pre-training the agent.
We show robust performance on the Real-Word RL benchmark, hinting at resiliency to environment perturbations during adaptation.
arXiv Detail & Related papers (2022-09-24T14:22:29Z) - Unified Automatic Control of Vehicular Systems with Reinforcement
Learning [64.63619662693068]
This article contributes a streamlined methodology for vehicular microsimulation.
It discovers high performance control strategies with minimal manual design.
The study reveals numerous emergent behaviors resembling wave mitigation, traffic signaling, and ramp metering.
arXiv Detail & Related papers (2022-07-30T16:23:45Z) - Reinforcement Learning for Mixed Autonomy Intersections [4.771833920251869]
We propose a model-free reinforcement learning method for controlling mixed autonomy traffic in simulated traffic networks.
Our method utilizes multi-agent policy decomposition which allows decentralized control based on local observations for an arbitrary number of controlled vehicles.
arXiv Detail & Related papers (2021-11-08T18:03:18Z) - A Deep Value-network Based Approach for Multi-Driver Order Dispatching [55.36656442934531]
We propose a deep reinforcement learning based solution for order dispatching.
We conduct large scale online A/B tests on DiDi's ride-dispatching platform.
Results show that CVNet consistently outperforms other recently proposed dispatching methods.
arXiv Detail & Related papers (2021-06-08T16:27:04Z) - Independent Reinforcement Learning for Weakly Cooperative Multiagent
Traffic Control Problem [22.733542222812158]
We use independent reinforcement learning (IRL) to solve a complex traffic cooperative control problem in this study.
To this, we model the traffic control problem as a partially observable weak cooperative traffic model (PO-WCTM) to optimize the overall traffic situation of a group of intersections.
Experimental results show that CIL-DDQN outperforms other methods in almost all performance indicators of the traffic control problem.
arXiv Detail & Related papers (2021-04-22T07:55:46Z) - Federated Learning on the Road: Autonomous Controller Design for
Connected and Autonomous Vehicles [109.71532364079711]
A new federated learning (FL) framework is proposed for designing the autonomous controller of connected and autonomous vehicles (CAVs)
A novel dynamic federated proximal (DFP) algorithm is proposed that accounts for the mobility of CAVs, the wireless fading channels, and the unbalanced and nonindependent and identically distributed data across CAVs.
A rigorous convergence analysis is performed for the proposed algorithm to identify how fast the CAVs converge to using the optimal controller.
arXiv Detail & Related papers (2021-02-05T19:57:47Z) - Leveraging the Capabilities of Connected and Autonomous Vehicles and
Multi-Agent Reinforcement Learning to Mitigate Highway Bottleneck Congestion [2.0010674945048468]
We present an RL-based multi-agent CAV control model to operate in mixed traffic.
The results suggest that even at CAV percent share of corridor traffic as low as 10%, CAVs can significantly mitigate bottlenecks in highway traffic.
arXiv Detail & Related papers (2020-10-12T03:52:10Z) - Reinforcement Learning Based Vehicle-cell Association Algorithm for
Highly Mobile Millimeter Wave Communication [53.47785498477648]
This paper investigates the problem of vehicle-cell association in millimeter wave (mmWave) communication networks.
We first formulate the user state (VU) problem as a discrete non-vehicle association optimization problem.
The proposed solution achieves up to 15% gains in terms sum of user complexity and 20% reduction in VUE compared to several baseline designs.
arXiv Detail & Related papers (2020-01-22T08:51:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.