Deep Reinforcement Learning based Dynamic Optimization of Bus Timetable
- URL: http://arxiv.org/abs/2107.07066v1
- Date: Thu, 15 Jul 2021 01:22:49 GMT
- Title: Deep Reinforcement Learning based Dynamic Optimization of Bus Timetable
- Authors: Guanqun Ai, Xingquan Zuo, Gang chen, and Binglin Wu
- Abstract summary: We propose a Deep Reinforcement Learning based bus Timetable dynamic Optimization method (DRL-TO)
A Deep Q-Network (DQN) is employed as the decision model to determine whether to dispatch a bus service during each minute of the service period.
DRL-TO can dynamically determine the departure intervals based on the real-time passenger flow, saving 8$%$ of vehicles and reducing 17$%$ of passengers' waiting time on average.
- Score: 4.337939117851783
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bus timetable optimization is a key issue to reduce operational cost of bus
companies and improve the service quality. Existing methods use exact or
heuristic algorithms to optimize the timetable in an offline manner. In
practice, the passenger flow may change significantly over time. Timetables
determined in offline cannot adjust the departure interval to satisfy the
changed passenger flow. Aiming at improving the online performance of bus
timetable, we propose a Deep Reinforcement Learning based bus Timetable dynamic
Optimization method (DRL-TO). In this method, the timetable optimization is
considered as a sequential decision problem. A Deep Q-Network (DQN) is employed
as the decision model to determine whether to dispatch a bus service during
each minute of the service period. Therefore, the departure intervals of bus
services are determined in real time in accordance with passenger demand. We
identify several new and useful state features for the DQN, including the load
factor, carrying capacity utilization rate, and the number of stranding
passengers. Taking into account both the interests of the bus company and
passengers, a reward function is designed, which includes the indicators of
full load rate, empty load rate, passengers' waiting time, and the number of
stranding passengers. Building on an existing method for calculating the
carrying capacity, we develop a new technique to enhance the matching degree at
each bus station. Experiments demonstrate that compared with the timetable
generated by the state-of-the-art bus timetable optimization approach based on
a memetic algorithm (BTOA-MA), Genetic Algorithm (GA) and the manual method,
DRL-TO can dynamically determine the departure intervals based on the real-time
passenger flow, saving 8$\%$ of vehicles and reducing 17$\%$ of passengers'
waiting time on average.
Related papers
- Rethinking Optimal Transport in Offline Reinforcement Learning [64.56896902186126]
In offline reinforcement learning, the data is provided by various experts and some of them can be sub-optimal.
To extract an efficient policy, it is necessary to emphstitch the best behaviors from the dataset.
We present an algorithm that aims to find a policy that maps states to a emphpartial distribution of the best expert actions for each given state.
arXiv Detail & Related papers (2024-10-17T22:36:43Z) - RL-MSA: a Reinforcement Learning-based Multi-line bus Scheduling
Approach [0.0]
Existing approaches typically generate a bus scheduling scheme in an offline manner and then schedule buses according to the scheme.
In this paper, MLBSP is modeled as a Markov Decision Process (MDP)
A Reinforcement Learning-based Multi-line bus Scheduling Approach (RL-MSA) is proposed for bus scheduling at both the offline and online phases.
arXiv Detail & Related papers (2024-03-11T07:07:05Z) - Short Run Transit Route Planning Decision Support System Using a Deep
Learning-Based Weighted Graph [0.0]
We propose a novel deep learning-based methodology for a decision support system that enables public transport planners to identify short-term route improvements rapidly.
By seamlessly adjusting specific sections of routes between two stops during specific times of the day, our method effectively reduces times and enhances PT services.
Using self-supervision, we train a deep learning model for predicting lateness values for road segments. These lateness values are then utilized as edge weights in the transportation graph, enabling efficient path searching.
arXiv Detail & Related papers (2023-08-24T14:37:55Z) - Deployment of Leader-Follower Automated Vehicle Systems for Smart Work
Zone Applications with a Queuing-based Traffic Assignment Approach [1.0355894890759376]
This manuscript focuses on optimizing the routing for ATMA vehicles in a network to minimize the system cost associated with the slow-moving operation.
A queuing-based traffic assignment approach is proposed to identify the system cost caused by the ATMA system.
The methodology is validated using a small-size and a large-size network and compared with two benchmark models to analyze the benefit of capacity drop modeling and QBTD travel time function.
arXiv Detail & Related papers (2023-07-23T16:35:05Z) - Real-Time Bus Arrival Prediction: A Deep Learning Approach for Enhanced
Urban Mobility [2.1374208474242815]
A prevalent challenge is the mismatch between actual bus arrival times and their scheduled counterparts, leading to disruptions in fixed schedules.
This research introduces an innovative, AI-based, data-driven methodology for predicting bus arrival times at various transit points (stations)
Through the deployment of a fully connected neural network, our method elevates the accuracy and efficiency of public bus transit systems.
arXiv Detail & Related papers (2023-03-27T16:45:22Z) - A Deep-Learning Based Optimization Approach to Address Stop-Skipping
Strategy in Urban Rail Transit Lines [0.0]
We introduce an advanced data-driven optimization approach to determine the optimal stop-skip pattern in urban rail transit lines.
We employ a Long Short-Term Memory (LSTM) deep learning model to predict the station-level demand rates for the peak hour.
Considering the exponential nature of the problem, we propose an Ant Colony Optimization technique to solve the problem in a desirable amount of time.
arXiv Detail & Related papers (2021-09-17T23:52:19Z) - A Deep Value-network Based Approach for Multi-Driver Order Dispatching [55.36656442934531]
We propose a deep reinforcement learning based solution for order dispatching.
We conduct large scale online A/B tests on DiDi's ride-dispatching platform.
Results show that CVNet consistently outperforms other recently proposed dispatching methods.
arXiv Detail & Related papers (2021-06-08T16:27:04Z) - Delay-Tolerant Constrained OCO with Application to Network Resource
Allocation [44.67787270821051]
We consider online convex optimization (OCO) with multi-slot feedback delay.
An agent makes a sequence of online decisions to minimize the accumulation of time-varying convex loss functions.
We propose Delay-Tolerant Constrained-OCO, which uses a novel constraint penalty with double regularization to tackle the asynchrony between information feedback and decision updates.
arXiv Detail & Related papers (2021-05-09T19:32:33Z) - Tracking Performance of Online Stochastic Learners [57.14673504239551]
Online algorithms are popular in large-scale learning settings due to their ability to compute updates on the fly, without the need to store and process data in large batches.
When a constant step-size is used, these algorithms also have the ability to adapt to drifts in problem parameters, such as data or model properties, and track the optimal solution with reasonable accuracy.
We establish a link between steady-state performance derived under stationarity assumptions and the tracking performance of online learners under random walk models.
arXiv Detail & Related papers (2020-04-04T14:16:27Z) - DADA: Differentiable Automatic Data Augmentation [58.560309490774976]
We propose Differentiable Automatic Data Augmentation (DADA) which dramatically reduces the cost.
We conduct extensive experiments on CIFAR-10, CIFAR-100, SVHN, and ImageNet datasets.
Results show our DADA is at least one order of magnitude faster than the state-of-the-art while achieving very comparable accuracy.
arXiv Detail & Related papers (2020-03-08T13:23:14Z) - Stacked Auto Encoder Based Deep Reinforcement Learning for Online
Resource Scheduling in Large-Scale MEC Networks [44.40722828581203]
An online resource scheduling framework is proposed for minimizing the sum of weighted task latency for all the Internet of things (IoT) users.
A deep reinforcement learning (DRL) based solution is proposed, which includes the following components.
A preserved and prioritized experience replay (2p-ER) is introduced to assist the DRL to train the policy network and find the optimal offloading policy.
arXiv Detail & Related papers (2020-01-24T23:01:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.