Timing the Match: A Deep Reinforcement Learning Approach for Ride-Hailing and Ride-Pooling Services
- URL: http://arxiv.org/abs/2503.13200v1
- Date: Mon, 17 Mar 2025 14:07:58 GMT
- Title: Timing the Match: A Deep Reinforcement Learning Approach for Ride-Hailing and Ride-Pooling Services
- Authors: Yiman Bao, Jie Gao, Jinke He, Frans A. Oliehoek, Oded Cats,
- Abstract summary: We propose an adaptive ride-matching strategy using deep reinforcement learning (RL) to determine when to perform matches based on real-time system conditions.<n>Our method continuously evaluates system states and executes matching at moments that minimize total passenger wait time.
- Score: 17.143444035884386
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Efficient timing in ride-matching is crucial for improving the performance of ride-hailing and ride-pooling services, as it determines the number of drivers and passengers considered in each matching process. Traditional batched matching methods often use fixed time intervals to accumulate ride requests before assigning matches. While this approach increases the number of available drivers and passengers for matching, it fails to adapt to real-time supply-demand fluctuations, often leading to longer passenger wait times and driver idle periods. To address this limitation, we propose an adaptive ride-matching strategy using deep reinforcement learning (RL) to dynamically determine when to perform matches based on real-time system conditions. Unlike fixed-interval approaches, our method continuously evaluates system states and executes matching at moments that minimize total passenger wait time. Additionally, we incorporate a potential-based reward shaping (PBRS) mechanism to mitigate sparse rewards, accelerating RL training and improving decision quality. Extensive empirical evaluations using a realistic simulator trained on real-world data demonstrate that our approach outperforms fixed-interval matching strategies, significantly reducing passenger waiting times and detour delays, thereby enhancing the overall efficiency of ride-hailing and ride-pooling systems.
Related papers
- Ride-Sourcing Vehicle Rebalancing with Service Accessibility Guarantees via Constrained Mean-Field Reinforcement Learning [42.070187224580344]
Rapid expansion of services such as Uber, Lyft and Didi Chuxing has reshaped urban transportation by offering flexible, on-demand mobility via mobile applications.
Inadequate rebalancing results in prolonged rider waiting times, inefficient vehicle utilization, and inequitable distribution services.
We introduce continuous-state mean-field control (MFC) and reinforcement learning (MFRL) models that explicitly represent each vehicle's precise location and employ continuous repositioning actions guided by the distribution of other vehicles.
arXiv Detail & Related papers (2025-03-31T15:00:11Z) - Spatio-temporal Incentives Optimization for Ride-hailing Services with
Offline Deep Reinforcement Learning [7.668735431419396]
We propose an offline reinforcement learning based method on the demand side to improve the utilization of transportation resources and customer satisfaction.
We adopt a deep-temporal learning method to learn the value of different time and location, then incentivize the ride requests passengers adjust the distribution of demand to balance the supply and demand in the system.
arXiv Detail & Related papers (2022-11-06T23:54:38Z) - Deep Explicit Duration Switching Models for Time Series [84.33678003781908]
We propose a flexible model that is capable of identifying both state- and time-dependent switching dynamics.
State-dependent switching is enabled by a recurrent state-to-switch connection.
An explicit duration count variable is used to improve the time-dependent switching behavior.
arXiv Detail & Related papers (2021-10-26T17:35:21Z) - Deep Reinforcement Learning based Dynamic Optimization of Bus Timetable [4.337939117851783]
We propose a Deep Reinforcement Learning based bus Timetable dynamic Optimization method (DRL-TO)
A Deep Q-Network (DQN) is employed as the decision model to determine whether to dispatch a bus service during each minute of the service period.
DRL-TO can dynamically determine the departure intervals based on the real-time passenger flow, saving 8$%$ of vehicles and reducing 17$%$ of passengers' waiting time on average.
arXiv Detail & Related papers (2021-07-15T01:22:49Z) - A Deep Value-network Based Approach for Multi-Driver Order Dispatching [55.36656442934531]
We propose a deep reinforcement learning based solution for order dispatching.
We conduct large scale online A/B tests on DiDi's ride-dispatching platform.
Results show that CVNet consistently outperforms other recently proposed dispatching methods.
arXiv Detail & Related papers (2021-06-08T16:27:04Z) - Improving Long-Term Metrics in Recommendation Systems using
Short-Horizon Offline RL [56.20835219296896]
We study session-based recommendation scenarios where we want to recommend items to users during sequential interactions to improve their long-term utility.
We develop a new batch RL algorithm called Short Horizon Policy Improvement (SHPI) that approximates policy-induced distribution shifts across sessions.
arXiv Detail & Related papers (2021-06-01T15:58:05Z) - Value Function is All You Need: A Unified Learning Framework for Ride
Hailing Platforms [57.21078336887961]
Large ride-hailing platforms, such as DiDi, Uber and Lyft, connect tens of thousands of vehicles in a city to millions of ride demands throughout the day.
We propose a unified value-based dynamic learning framework (V1D3) for tackling both tasks.
arXiv Detail & Related papers (2021-05-18T19:22:24Z) - Real-world Ride-hailing Vehicle Repositioning using Deep Reinforcement
Learning [52.2663102239029]
We present a new practical framework based on deep reinforcement learning and decision-time planning for real-world vehicle on idle-hailing platforms.
Our approach learns ride-based state-value function using a batch training algorithm with deep value.
We benchmark our algorithm with baselines in a ride-hailing simulation environment to demonstrate its superiority in improving income efficiency.
arXiv Detail & Related papers (2021-03-08T05:34:05Z) - Equilibrium Inverse Reinforcement Learning for Ride-hailing Vehicle
Network [1.599072005190786]
We formulate the problem of passenger-vehicle matching in a sparsely connected graph.
We propose an algorithm to derive an equilibrium policy in a multi-agent environment.
arXiv Detail & Related papers (2021-02-13T03:18:44Z) - Vehicular Cooperative Perception Through Action Branching and Federated
Reinforcement Learning [101.64598586454571]
A novel framework is proposed to allow reinforcement learning-based vehicular association, resource block (RB) allocation, and content selection of cooperative perception messages (CPMs)
A federated RL approach is introduced in order to speed up the training process across vehicles.
Results show that federated RL improves the training process, where better policies can be achieved within the same amount of time compared to the non-federated approach.
arXiv Detail & Related papers (2020-12-07T02:09:15Z) - Zone pAth Construction (ZAC) based Approaches for Effective Real-Time
Ridesharing [30.964687022746226]
Key challenge in real-time ridesharing systems is to group the "right" requests to travel together in the "right" available vehicles in real-time.
We contribute both myopic (ridesharing assignment focussed on current requests only) and non-myopic (ridesharing considers impact on expected future requests) approaches that employ zone paths.
arXiv Detail & Related papers (2020-09-13T17:57:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.