Mutual Information as Intrinsic Reward of Reinforcement Learning Agents
for On-demand Ride Pooling
- URL: http://arxiv.org/abs/2312.15195v2
- Date: Sun, 7 Jan 2024 12:12:39 GMT
- Title: Mutual Information as Intrinsic Reward of Reinforcement Learning Agents
for On-demand Ride Pooling
- Authors: Xianjie Zhang, Jiahao Sun, Chen Gong, Kai Wang, Yifei Cao, Hao Chen,
Hao Chen, Yu Liu
- Abstract summary: On-demand vehicle pooling services allow each vehicle to serve multiple passengers at a time.
Existing algorithms often only consider revenue, which makes it difficult for requests with unusual distribution to get a ride.
We propose a framework for dispatching for ride pooling tasks, which splits the city into discrete dispatching and uses the reinforcement learning (RL) algorithm to dispatch vehicles in these regions.
- Score: 19.247162142334076
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The emergence of on-demand ride pooling services allows each vehicle to serve
multiple passengers at a time, thus increasing drivers' income and enabling
passengers to travel at lower prices than taxi/car on-demand services (only one
passenger can be assigned to a car at a time like UberX and Lyft). Although
on-demand ride pooling services can bring so many benefits, ride pooling
services need a well-defined matching strategy to maximize the benefits for all
parties (passengers, drivers, aggregation companies and environment), in which
the regional dispatching of vehicles has a significant impact on the matching
and revenue. Existing algorithms often only consider revenue maximization,
which makes it difficult for requests with unusual distribution to get a ride.
How to increase revenue while ensuring a reasonable assignment of requests
brings a challenge to ride pooling service companies (aggregation companies).
In this paper, we propose a framework for vehicle dispatching for ride pooling
tasks, which splits the city into discrete dispatching regions and uses the
reinforcement learning (RL) algorithm to dispatch vehicles in these regions. We
also consider the mutual information (MI) between vehicle and order
distribution as the intrinsic reward of the RL algorithm to improve the
correlation between their distributions, thus ensuring the possibility of
getting a ride for unusually distributed requests. In experimental results on a
real-world taxi dataset, we demonstrate that our framework can significantly
increase revenue up to an average of 3\% over the existing best on-demand ride
pooling method.
Related papers
- GPT-Augmented Reinforcement Learning with Intelligent Control for Vehicle Dispatching [82.19172267487998]
GARLIC: a framework of GPT-Augmented Reinforcement Learning with Intelligent Control for vehicle dispatching.
This paper introduces GARLIC: a framework of GPT-Augmented Reinforcement Learning with Intelligent Control for vehicle dispatching.
arXiv Detail & Related papers (2024-08-19T08:23:38Z) - Fair collaborative vehicle routing: A deep multi-agent reinforcement
learning approach [49.00137468773683]
Collaborative vehicle routing occurs when carriers collaborate through sharing their transportation requests and performing transportation requests on behalf of each other.
Traditional game theoretic solution concepts are expensive to calculate as the characteristic function scales exponentially with the number of agents.
We propose to model this problem as a coalitional bargaining game solved using deep multi-agent reinforcement learning.
arXiv Detail & Related papers (2023-10-26T15:42:29Z) - Coalitional Bargaining via Reinforcement Learning: An Application to
Collaborative Vehicle Routing [49.00137468773683]
Collaborative Vehicle Routing is where delivery companies cooperate by sharing their delivery information and performing delivery requests on behalf of each other.
This achieves economies of scale and thus reduces cost, greenhouse gas emissions, and road congestion.
But which company should partner with whom, and how much should each company be compensated?
Traditional game theoretic solution concepts, such as the Shapley value or nucleolus, are difficult to calculate for the real-world problem of Collaborative Vehicle Routing.
arXiv Detail & Related papers (2023-10-26T15:04:23Z) - Safe Model-Based Multi-Agent Mean-Field Reinforcement Learning [48.667697255912614]
Mean-field reinforcement learning addresses the policy of a representative agent interacting with the infinite population of identical agents.
We propose Safe-M$3$-UCRL, the first model-based mean-field reinforcement learning algorithm that attains safe policies even in the case of unknown transitions.
Our algorithm effectively meets the demand in critical areas while ensuring service accessibility in regions with low demand.
arXiv Detail & Related papers (2023-06-29T15:57:07Z) - Value Function is All You Need: A Unified Learning Framework for Ride
Hailing Platforms [57.21078336887961]
Large ride-hailing platforms, such as DiDi, Uber and Lyft, connect tens of thousands of vehicles in a city to millions of ride demands throughout the day.
We propose a unified value-based dynamic learning framework (V1D3) for tackling both tasks.
arXiv Detail & Related papers (2021-05-18T19:22:24Z) - Vehicular Cooperative Perception Through Action Branching and Federated
Reinforcement Learning [101.64598586454571]
A novel framework is proposed to allow reinforcement learning-based vehicular association, resource block (RB) allocation, and content selection of cooperative perception messages (CPMs)
A federated RL approach is introduced in order to speed up the training process across vehicles.
Results show that federated RL improves the training process, where better policies can be achieved within the same amount of time compared to the non-federated approach.
arXiv Detail & Related papers (2020-12-07T02:09:15Z) - PassGoodPool: Joint Passengers and Goods Fleet Management with
Reinforcement Learning aided Pricing, Matching, and Route Planning [29.73314892749729]
We present a demand aware fleet management framework for combined goods and passenger transportation.
Our proposed model is deployable independently within each vehicle as this minimizes computational costs associated with the growth of distributed systems.
arXiv Detail & Related papers (2020-11-17T23:15:03Z) - A Distributed Model-Free Ride-Sharing Approach for Joint Matching,
Pricing, and Dispatching using Deep Reinforcement Learning [32.0512015286512]
We present a dynamic, demand aware, and pricing-based vehicle-passenger matching and route planning framework.
Our framework is validated using the New York City Taxi dataset.
Experimental results show the effectiveness of our approach in real-time and large scale settings.
arXiv Detail & Related papers (2020-10-05T03:13:47Z) - Competitive Ratios for Online Multi-capacity Ridesharing [30.964687022746226]
In multi-capacity ridesharing, multiple requests (e.g., customers, food items, parcels) with different origin and destination pairs travel in one resource.
Online multi-capacity ridesharing is extremely challenging as the underlying matching graph is no longer bipartite.
This paper presents the first approach with bounds on the competitive ratio for online multi-capacity ridesharing.
arXiv Detail & Related papers (2020-09-16T20:29:21Z) - Zone pAth Construction (ZAC) based Approaches for Effective Real-Time
Ridesharing [30.964687022746226]
Key challenge in real-time ridesharing systems is to group the "right" requests to travel together in the "right" available vehicles in real-time.
We contribute both myopic (ridesharing assignment focussed on current requests only) and non-myopic (ridesharing considers impact on expected future requests) approaches that employ zone paths.
arXiv Detail & Related papers (2020-09-13T17:57:15Z) - Balancing Taxi Distribution in A City-Scale Dynamic Ridesharing Service:
A Hybrid Solution Based on Demand Learning [0.0]
We study the challenging problem of how to balance taxi distribution across a city in a dynamic ridesharing service.
We propose a hybrid solution involving a series of algorithms: the Correlated Pooling collects correlated rider requests, the Adjacency Ride-Matching based on Demand Learning assigns taxis to riders, and the Greedy Idle Movement aims to direct taxis without a current assignment to the areas with riders in need of service.
arXiv Detail & Related papers (2020-07-27T07:08:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.