Reinforcement Learning for Freight Booking Control Problems
- URL: http://arxiv.org/abs/2102.00092v3
- Date: Wed, 5 Apr 2023 00:39:18 GMT
- Title: Reinforcement Learning for Freight Booking Control Problems
- Authors: Justin Dumouchelle, Emma Frejinger, Andrea Lodi
- Abstract summary: Booking control problems are sequential decision-making problems in revenue management.
We train a supervised learning model to predict the objective of an operational problem.
We then deploy the model within reinforcement learning algorithms to compute control policies.
- Score: 5.08128537391027
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Booking control problems are sequential decision-making problems that occur
in the domain of revenue management. More precisely, freight booking control
focuses on the problem of deciding to accept or reject bookings: given a
limited capacity, accept a booking request or reject it to reserve capacity for
future bookings with potentially higher revenue. This problem can be formulated
as a finite-horizon stochastic dynamic program, where accepting a set of
requests results in a profit at the end of the booking period that depends on
the cost of fulfilling the accepted bookings. For many freight applications,
the cost of fulfilling requests is obtained by solving an operational
decision-making problem, which often requires the solutions to mixed-integer
linear programs. Routinely solving such operational problems when deploying
reinforcement learning algorithms may be too time consuming. The majority of
booking control policies are obtained by solving problem-specific mathematical
programming relaxations that are often non-trivial to generalize to new
problems and, in some cases, provide quite crude approximations.
In this work, we propose a two-phase approach: we first train a supervised
learning model to predict the objective of the operational problem, and then we
deploy the model within reinforcement learning algorithms to compute control
policies. This approach is general: it can be used every time the objective
function of the end-of-horizon operational problem can be predicted, and it is
particularly suitable to those cases where such problems are computationally
hard. Furthermore, it allows one to leverage the recent advances in
reinforcement learning as routinely solving the operational problem is replaced
with a single prediction. Our methodology is evaluated on two booking control
problems in the literature, namely, distributional logistics and airline cargo
management.
Related papers
- Zero-shot Generalization in Inventory Management: Train, then Estimate and Decide [0.0]
Deploying deep reinforcement learning (DRL) in real-world inventory management presents challenges.
These challenges highlight a research gap, suggesting a need for a unifying framework to model and solve sequential decision-making under parameter uncertainty.
We address this by exploring an underexplored area of DRL for inventory management: training generally capable agents (GCAs) under zero-shot generalization (ZSG)
arXiv Detail & Related papers (2024-11-01T11:20:05Z) - Contractual Reinforcement Learning: Pulling Arms with Invisible Hands [68.77645200579181]
We propose a theoretical framework for aligning economic interests of different stakeholders in the online learning problems through contract design.
For the planning problem, we design an efficient dynamic programming algorithm to determine the optimal contracts against the far-sighted agent.
For the learning problem, we introduce a generic design of no-regret learning algorithms to untangle the challenges from robust design of contracts to the balance of exploration and exploitation.
arXiv Detail & Related papers (2024-07-01T16:53:00Z) - Predicting Probabilities of Error to Combine Quantization and Early Exiting: QuEE [68.6018458996143]
We propose a more general dynamic network that can combine both quantization and early exit dynamic network: QuEE.
Our algorithm can be seen as a form of soft early exiting or input-dependent compression.
The crucial factor of our approach is accurate prediction of the potential accuracy improvement achievable through further computation.
arXiv Detail & Related papers (2024-06-20T15:25:13Z) - Learning with Posterior Sampling for Revenue Management under Time-varying Demand [36.22276574805786]
We discuss the revenue management problem to maximize revenue by pricing items or services.
One challenge in this problem is that the demand distribution is unknown and varies over time in real applications such as airline and retail industries.
arXiv Detail & Related papers (2024-05-08T09:28:26Z) - Model-Based Reinforcement Learning Control of Reaction-Diffusion
Problems [0.0]
reinforcement learning has been applied to decision-making in several applications, most notably in games.
We introduce two novel reward functions to drive the flow of the transported field.
Results show that certain controls can be implemented successfully in these applications.
arXiv Detail & Related papers (2024-02-22T11:06:07Z) - Successive Refinement in Large-Scale Computation: Advancing Model
Inference Applications [67.76749044675721]
We introduce solutions for layered-resolution computation.
These solutions allow lower-resolution results to be obtained at an earlier stage than the final result.
arXiv Detail & Related papers (2024-02-11T15:36:33Z) - Attention-based Reinforcement Learning for Combinatorial Optimization: Application to Job Shop Scheduling Problem [2.024210754085351]
This study proposes an innovative attention-based reinforcement learning method specifically designed for the category of job shop scheduling problems.
A key finding of this research is the ability of our trained learners within the proposed method to be repurposed for larger-scale problems that were not part of the initial training set.
arXiv Detail & Related papers (2024-01-29T21:31:54Z) - An End-to-End Reinforcement Learning Approach for Job-Shop Scheduling
Problems Based on Constraint Programming [5.070542698701157]
This paper proposes a novel end-to-end approach to solving scheduling problems by means of CP and Reinforcement Learning (RL)
Our approach leverages existing CP solvers to train an agent learning a Priority Dispatching Rule (PDR) that generalizes well to large instances, even from separate datasets.
arXiv Detail & Related papers (2023-06-09T08:24:56Z) - A Universal Error Measure for Input Predictions Applied to Online Graph
Problems [57.58926849872494]
We introduce a novel measure for quantifying the error in input predictions.
The measure captures errors due to absent predicted requests as well as unpredicted actual requests.
arXiv Detail & Related papers (2022-05-25T15:24:03Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Probably Approximately Correct Constrained Learning [135.48447120228658]
We develop a generalization theory based on the probably approximately correct (PAC) learning framework.
We show that imposing a learner does not make a learning problem harder in the sense that any PAC learnable class is also a constrained learner.
We analyze the properties of this solution and use it to illustrate how constrained learning can address problems in fair and robust classification.
arXiv Detail & Related papers (2020-06-09T19:59:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.