Spatio-temporal Incentives Optimization for Ride-hailing Services with
Offline Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2211.03240v1
- Date: Sun, 6 Nov 2022 23:54:38 GMT
- Title: Spatio-temporal Incentives Optimization for Ride-hailing Services with
Offline Deep Reinforcement Learning
- Authors: Yanqiu Wu, Qingyang Li, Zhiwei Qin
- Abstract summary: We propose an offline reinforcement learning based method on the demand side to improve the utilization of transportation resources and customer satisfaction.
We adopt a deep-temporal learning method to learn the value of different time and location, then incentivize the ride requests passengers adjust the distribution of demand to balance the supply and demand in the system.
- Score: 7.668735431419396
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A fundamental question in any peer-to-peer ride-sharing system is how to,
both effectively and efficiently, meet the request of passengers to balance the
supply and demand in real time. On the passenger side, traditional approaches
focus on pricing strategies by increasing the probability of users' call to
adjust the distribution of demand. However, previous methods do not take into
account the impact of changes in strategy on future supply and demand changes,
which means drivers are repositioned to different destinations due to
passengers' calls, which will affect the driver's income for a period of time
in the future. Motivated by this observation, we make an attempt to optimize
the distribution of demand to handle this problem by learning the long-term
spatio-temporal values as a guideline for pricing strategy. In this study, we
propose an offline deep reinforcement learning based method focusing on the
demand side to improve the utilization of transportation resources and customer
satisfaction. We adopt a spatio-temporal learning method to learn the value of
different time and location, then incentivize the ride requests of passengers
to adjust the distribution of demand to balance the supply and demand in the
system. In particular, we model the problem as a Markov Decision Process (MDP).
Related papers
- Online Learning of Weakly Coupled MDP Policies for Load Balancing and Auto Scaling [42.6574685545681]
This paper introduces a novel model and algorithms for tuning load balancers coupled with auto scalers, considering bursty traffic arriving at finite queues.
We begin by presenting the problem as a weakly coupled Markov Decision Processes (MDP), solvable via a linear program (LP)
We extend it to tackle the problem of online parameter learning and policy optimization using a two-timescale algorithm based on the LP Lagrangian.
arXiv Detail & Related papers (2024-06-20T09:34:24Z) - Fair collaborative vehicle routing: A deep multi-agent reinforcement
learning approach [49.00137468773683]
Collaborative vehicle routing occurs when carriers collaborate through sharing their transportation requests and performing transportation requests on behalf of each other.
Traditional game theoretic solution concepts are expensive to calculate as the characteristic function scales exponentially with the number of agents.
We propose to model this problem as a coalitional bargaining game solved using deep multi-agent reinforcement learning.
arXiv Detail & Related papers (2023-10-26T15:42:29Z) - Coalitional Bargaining via Reinforcement Learning: An Application to
Collaborative Vehicle Routing [49.00137468773683]
Collaborative Vehicle Routing is where delivery companies cooperate by sharing their delivery information and performing delivery requests on behalf of each other.
This achieves economies of scale and thus reduces cost, greenhouse gas emissions, and road congestion.
But which company should partner with whom, and how much should each company be compensated?
Traditional game theoretic solution concepts, such as the Shapley value or nucleolus, are difficult to calculate for the real-world problem of Collaborative Vehicle Routing.
arXiv Detail & Related papers (2023-10-26T15:04:23Z) - Optimizing Credit Limit Adjustments Under Adversarial Goals Using
Reinforcement Learning [42.303733194571905]
We seek to find and automatize an optimal credit card limit adjustment policy by employing reinforcement learning techniques.
Our research establishes a conceptual structure for applying reinforcement learning framework to credit limit adjustment.
arXiv Detail & Related papers (2023-06-27T16:10:36Z) - STEF-DHNet: Spatiotemporal External Factors Based Deep Hybrid Network
for Enhanced Long-Term Taxi Demand Prediction [16.07685260834701]
This paper introduces STEF-DHNet, a demand prediction model that integrates external features astemporal information.
It is evaluated using a long-term performance metric called the rolling error, which assesses its ability to maintain high accuracy over long periods without retraining.
The results show that STEF-DHNet outperforms existing state-of-the-art methods on three diverse datasets.
arXiv Detail & Related papers (2023-06-26T07:37:50Z) - Structured Dynamic Pricing: Optimal Regret in a Global Shrinkage Model [50.06663781566795]
We consider a dynamic model with the consumers' preferences as well as price sensitivity varying over time.
We measure the performance of a dynamic pricing policy via regret, which is the expected revenue loss compared to a clairvoyant that knows the sequence of model parameters in advance.
Our regret analysis results not only demonstrate optimality of the proposed policy but also show that for policy planning it is essential to incorporate available structural information.
arXiv Detail & Related papers (2023-03-28T00:23:23Z) - Multiagent Reinforcement Learning for Autonomous Routing and Pickup
Problem with Adaptation to Variable Demand [1.8505047763172104]
We derive a learning framework to generate routing/pickup policies for a fleet of autonomous vehicles tasked with appearing requests on a city map.
We focus on policies that give rise to coordination amongst the vehicles, thereby reducing wait times for servicing requests.
We propose a mechanism for switching the originally trained offline approximation when the current demand is outside the original validity region.
arXiv Detail & Related papers (2022-11-28T01:11:11Z) - Meta-Learning over Time for Destination Prediction Tasks [53.12827614887103]
A need to understand and predict vehicles' behavior underlies both public and private goals in the transportation domain.
Recent studies have found, at best, only marginal improvements in predictive performance from incorporating temporal information.
We propose an approach based on hypernetworks, in which a neural network learns to change its own weights in response to an input.
arXiv Detail & Related papers (2022-06-29T17:58:12Z) - A Deep-Learning Based Optimization Approach to Address Stop-Skipping
Strategy in Urban Rail Transit Lines [0.0]
We introduce an advanced data-driven optimization approach to determine the optimal stop-skip pattern in urban rail transit lines.
We employ a Long Short-Term Memory (LSTM) deep learning model to predict the station-level demand rates for the peak hour.
Considering the exponential nature of the problem, we propose an Ant Colony Optimization technique to solve the problem in a desirable amount of time.
arXiv Detail & Related papers (2021-09-17T23:52:19Z) - A Distributed Model-Free Ride-Sharing Approach for Joint Matching,
Pricing, and Dispatching using Deep Reinforcement Learning [32.0512015286512]
We present a dynamic, demand aware, and pricing-based vehicle-passenger matching and route planning framework.
Our framework is validated using the New York City Taxi dataset.
Experimental results show the effectiveness of our approach in real-time and large scale settings.
arXiv Detail & Related papers (2020-10-05T03:13:47Z) - Learn to Earn: Enabling Coordination within a Ride Hailing Fleet [5.016829322655594]
We study the problem of optimizing social welfare objectives on multi sided ride hailing platforms such as Uber, Lyft, etc.
An ideal solution aims to minimize the response time for each hyper local passenger ride request, while simultaneously maintaining high demand satisfaction and supply utilization across the entire city.
arXiv Detail & Related papers (2020-06-19T00:20:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.