Flatland Competition 2020: MAPF and MARL for Efficient Train
Coordination on a Grid World
- URL: http://arxiv.org/abs/2103.16511v1
- Date: Tue, 30 Mar 2021 17:13:29 GMT
- Title: Flatland Competition 2020: MAPF and MARL for Efficient Train
Coordination on a Grid World
- Authors: Florian Laurent, Manuel Schneider, Christian Scheller, Jeremy Watson,
Jiaoyang Li, Zhe Chen, Yi Zheng, Shao-Hung Chan, Konstantin Makhnev, Oleg
Svidchenko, Vladimir Egorov, Dmitry Ivanov, Aleksei Shpilman, Evgenija
Spirovska, Oliver Tanevski, Aleksandar Nikov, Ramon Grunder, David Galevski,
Jakov Mitrovski, Guillaume Sartoretti, Zhiyao Luo, Mehul Damani, Nilabha
Bhattacharya, Shivam Agarwal, Adrian Egli, Erik Nygren, Sharada Mohanty
- Abstract summary: The Flatland competition aimed at finding novel approaches to solve the vehicle re-scheduling problem (VRSP)
The VRSP is concerned with scheduling trips in traffic networks and the re-scheduling of vehicles when disruptions occur.
The ever-growing complexity of modern railway networks makes dynamic real-time scheduling of traffic virtually impossible.
- Score: 49.80905654161763
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Flatland competition aimed at finding novel approaches to solve the
vehicle re-scheduling problem (VRSP). The VRSP is concerned with scheduling
trips in traffic networks and the re-scheduling of vehicles when disruptions
occur, for example the breakdown of a vehicle. While solving the VRSP in
various settings has been an active area in operations research (OR) for
decades, the ever-growing complexity of modern railway networks makes dynamic
real-time scheduling of traffic virtually impossible. Recently, multi-agent
reinforcement learning (MARL) has successfully tackled challenging tasks where
many agents need to be coordinated, such as multiplayer video games. However,
the coordination of hundreds of agents in a real-life setting like a railway
network remains challenging and the Flatland environment used for the
competition models these real-world properties in a simplified manner.
Submissions had to bring as many trains (agents) to their target stations in as
little time as possible. While the best submissions were in the OR category,
participants found many promising MARL approaches. Using both centralized and
decentralized learning based approaches, top submissions used graph
representations of the environment to construct tree-based observations.
Further, different coordination mechanisms were implemented, such as
communication and prioritization between agents. This paper presents the
competition setup, four outstanding solutions to the competition, and a
cross-comparison between them.
Related papers
- Cooperative Path Planning with Asynchronous Multiagent Reinforcement Learning [4.640948267127441]
shortest path problem (SPP) with multiple source-destination pairs (MSD)
In this paper, we study the shortest path problem (SPP) with multiple source-destination pairs (MSD), namely MSD-SPP, to minimize average travel time of all shortest paths.
arXiv Detail & Related papers (2024-09-01T15:48:14Z) - A Reinforcement Learning Approach for Dynamic Rebalancing in
Bike-Sharing System [11.237099288412558]
Bike-Sharing Systems provide eco-friendly urban mobility, contributing to the alleviation of traffic congestion and healthier lifestyles.
Devising effective rebalancing strategies using vehicles to redistribute bikes among stations is therefore of uttermost importance for operators.
This paper introduces atemporal reinforcement learning algorithm for the dynamic rebalancing problem with multiple vehicles.
arXiv Detail & Related papers (2024-02-05T23:46:42Z) - Optimizing Crowd-Aware Multi-Agent Path Finding through Local Communication with Graph Neural Networks [15.88107215224685]
Multi-Agent Path Finding (MAPF) in crowded environments presents a challenging problem in motion planning.
We introduce CRAMP, a novel crowd-aware decentralized reinforcement learning approach to address this problem.
CRAMP improves the solution quality up to 59% measured in makespan and collision count, and up to 35% improvement in success rate in comparison to previous methods.
arXiv Detail & Related papers (2023-09-19T03:02:43Z) - Learning to Sail Dynamic Networks: The MARLIN Reinforcement Learning
Framework for Congestion Control in Tactical Environments [53.08686495706487]
This paper proposes an RL framework that leverages an accurate and parallelizable emulation environment to reenact the conditions of a tactical network.
We evaluate our RL learning framework by training a MARLIN agent in conditions replicating a bottleneck link transition between a Satellite Communication (SATCOM) and an UHF Wide Band (UHF) radio link.
arXiv Detail & Related papers (2023-06-27T16:15:15Z) - NeurIPS 2022 Competition: Driving SMARTS [60.948652154552136]
Driving SMARTS is a regular competition designed to tackle problems caused by the distribution shift in dynamic interaction contexts.
The proposed competition supports methodologically diverse solutions, such as reinforcement learning (RL) and offline learning methods.
arXiv Detail & Related papers (2022-11-14T17:10:53Z) - Value Function is All You Need: A Unified Learning Framework for Ride
Hailing Platforms [57.21078336887961]
Large ride-hailing platforms, such as DiDi, Uber and Lyft, connect tens of thousands of vehicles in a city to millions of ride demands throughout the day.
We propose a unified value-based dynamic learning framework (V1D3) for tackling both tasks.
arXiv Detail & Related papers (2021-05-18T19:22:24Z) - Competing Adaptive Networks [56.56653763124104]
We develop an algorithm for decentralized competition among teams of adaptive agents.
We present an application in the decentralized training of generative adversarial neural networks.
arXiv Detail & Related papers (2021-03-29T14:42:15Z) - Multi-Agent Reinforcement Learning for Markov Routing Games: A New
Modeling Paradigm For Dynamic Traffic Assignment [11.093194714316434]
We develop a Markov routing game (MRG) in which each agent learns and updates her own en-route path choice policy.
We show that the routing behavior of intelligent agents is shown to converge to the classical notion of predictive dynamic user equilibrium.
arXiv Detail & Related papers (2020-11-22T02:31:14Z) - Multi-Agent Routing Value Iteration Network [88.38796921838203]
We propose a graph neural network based model that is able to perform multi-agent routing based on learned value in a sparsely connected graph.
We show that our model trained with only two agents on graphs with a maximum of 25 nodes can easily generalize to situations with more agents and/or nodes.
arXiv Detail & Related papers (2020-07-09T22:16:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.