Fleet Rebalancing for Expanding Shared e-Mobility Systems: A Multi-agent
Deep Reinforcement Learning Approach
- URL: http://arxiv.org/abs/2211.06136v1
- Date: Fri, 11 Nov 2022 11:25:30 GMT
- Title: Fleet Rebalancing for Expanding Shared e-Mobility Systems: A Multi-agent
Deep Reinforcement Learning Approach
- Authors: Man Luo, Bowen Du, Wenzhe Zhang, Tianyou Song, Kun Li, Hongming Zhu,
Mark Birkin, Hongkai Wen
- Abstract summary: A key challenge in the operation of shared e-mobility systems is fleet rebalancing.
We first investigate rich sets of data collected from a real-world shared e-mobility system for one year.
With the learned knowledge we design a high-fidelity simulator, which is able to abstract key operation details of EV sharing.
Then we model the rebalancing task for shared e-mobility systems under continuous expansion as a Multi-Agent Reinforcement Learning (MARL) problem.
- Score: 17.193480676611358
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The electrification of shared mobility has become popular across the globe.
Many cities have their new shared e-mobility systems deployed, with
continuously expanding coverage from central areas to the city edges. A key
challenge in the operation of these systems is fleet rebalancing, i.e., how EVs
should be repositioned to better satisfy future demand. This is particularly
challenging in the context of expanding systems, because i) the range of the
EVs is limited while charging time is typically long, which constrain the
viable rebalancing operations; and ii) the EV stations in the system are
dynamically changing, i.e., the legitimate targets for rebalancing operations
can vary over time. We tackle these challenges by first investigating rich sets
of data collected from a real-world shared e-mobility system for one year,
analyzing the operation model, usage patterns and expansion dynamics of this
new mobility mode. With the learned knowledge we design a high-fidelity
simulator, which is able to abstract key operation details of EV sharing at
fine granularity. Then we model the rebalancing task for shared e-mobility
systems under continuous expansion as a Multi-Agent Reinforcement Learning
(MARL) problem, which directly takes the range and charging properties of the
EVs into account. We further propose a novel policy optimization approach with
action cascading, which is able to cope with the expansion dynamics and solve
the formulated MARL. We evaluate the proposed approach extensively, and
experimental results show that our approach outperforms the state-of-the-art,
offering significant performance gain in both satisfied demand and net revenue.
Related papers
- A Reinforcement Learning Approach for Dynamic Rebalancing in
Bike-Sharing System [11.237099288412558]
Bike-Sharing Systems provide eco-friendly urban mobility, contributing to the alleviation of traffic congestion and healthier lifestyles.
Devising effective rebalancing strategies using vehicles to redistribute bikes among stations is therefore of uttermost importance for operators.
This paper introduces atemporal reinforcement learning algorithm for the dynamic rebalancing problem with multiple vehicles.
arXiv Detail & Related papers (2024-02-05T23:46:42Z) - HAZARD Challenge: Embodied Decision Making in Dynamically Changing
Environments [93.94020724735199]
HAZARD consists of three unexpected disaster scenarios, including fire, flood, and wind.
This benchmark enables us to evaluate autonomous agents' decision-making capabilities across various pipelines.
arXiv Detail & Related papers (2024-01-23T18:59:43Z) - Rethinking Urban Mobility Prediction: A Super-Multivariate Time Series
Forecasting Approach [71.67506068703314]
Long-term urban mobility predictions play a crucial role in the effective management of urban facilities and services.
Traditionally, urban mobility data has been structured as videos, treating longitude and latitude as fundamental pixels.
In our research, we introduce a fresh perspective on urban mobility prediction.
Instead of oversimplifying urban mobility data as traditional video data, we regard it as a complex time series.
arXiv Detail & Related papers (2023-12-04T07:39:05Z) - Model-Based Reinforcement Learning with Isolated Imaginations [61.67183143982074]
We propose Iso-Dream++, a model-based reinforcement learning approach.
We perform policy optimization based on the decoupled latent imaginations.
This enables long-horizon visuomotor control tasks to benefit from isolating mixed dynamics sources in the wild.
arXiv Detail & Related papers (2023-03-27T02:55:56Z) - Graph Attention Multi-Agent Fleet Autonomy for Advanced Air Mobility [7.533471021886634]
We introduce the concept of partially observable advanced air mobility games to coordinate a fleet of aerial vehicles.
We propose a novel graph attention-decoder (HetGAT Enc-Dec) network neural-based policy.
We show that the learned policy generalizes to various fleet compositions, demand patterns, and observation topologies.
arXiv Detail & Related papers (2023-02-14T20:48:00Z) - Towards More Efficient Shared Autonomous Mobility: A Learning-Based
Fleet Repositioning Approach [0.0]
This paper formulates SAMS fleet as a Markov Decision Process and presents a reinforcement learning-based repositioning (RLR) approach called integrated system-agent repositioning (ISR)
The ISR learns to respond to evolving demand patterns without explicit demand forecasting and to cooperate with optimization-based passenger-to-vehicle assignment.
Results demonstrate the RLR approaches' substantial reductions in passenger wait times, over 50%, relative to the JO approach.
arXiv Detail & Related papers (2022-10-16T23:30:46Z) - Learning energy-efficient driving behaviors by imitating experts [75.12960180185105]
This paper examines the role of imitation learning in bridging the gap between control strategies and realistic limitations in communication and sensing.
We show that imitation learning can succeed in deriving policies that, if adopted by 5% of vehicles, may boost the energy-efficiency of networks with varying traffic conditions by 15% using only local observations.
arXiv Detail & Related papers (2022-06-28T17:08:31Z) - Event-Aware Multimodal Mobility Nowcasting [11.540605108140538]
Events-awaretemporal network EAST-Net is evaluated on real-world datasets with a wide variety and coverage of societal datasets.
Results verify the superiority of our approach compared with the state-of-the-art baselines.
arXiv Detail & Related papers (2021-12-14T12:35:20Z) - A Modular and Transferable Reinforcement Learning Framework for the
Fleet Rebalancing Problem [2.299872239734834]
We propose a modular framework for fleet rebalancing based on model-free reinforcement learning (RL)
We formulate RL state and action spaces as distributions over a grid of the operating area, making the framework scalable.
Numerical experiments, using real-world trip and network data, demonstrate that this approach has several distinct advantages over baseline methods.
arXiv Detail & Related papers (2021-05-27T16:32:28Z) - Value Function is All You Need: A Unified Learning Framework for Ride
Hailing Platforms [57.21078336887961]
Large ride-hailing platforms, such as DiDi, Uber and Lyft, connect tens of thousands of vehicles in a city to millions of ride demands throughout the day.
We propose a unified value-based dynamic learning framework (V1D3) for tackling both tasks.
arXiv Detail & Related papers (2021-05-18T19:22:24Z) - GEM: Group Enhanced Model for Learning Dynamical Control Systems [78.56159072162103]
We build effective dynamical models that are amenable to sample-based learning.
We show that learning the dynamics on a Lie algebra vector space is more effective than learning a direct state transition model.
This work sheds light on a connection between learning of dynamics and Lie group properties, which opens doors for new research directions.
arXiv Detail & Related papers (2021-04-07T01:08:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.