Network Revenue Management with Demand Learning and Fair
Resource-Consumption Balancing
- URL: http://arxiv.org/abs/2207.11159v3
- Date: Fri, 8 Sep 2023 03:23:37 GMT
- Title: Network Revenue Management with Demand Learning and Fair
Resource-Consumption Balancing
- Authors: Xi Chen, Jiameng Lyu, Yining Wang, Yuan Zhou
- Abstract summary: We study the price-based network revenue management (NRM) problem with both demand learning and fair resource-consumption balancing.
We propose a primal-dual-type online policy with the Upper-Confidence-Bound (UCB) demand learning method to maximize the regularized revenue.
Our algorithm achieves a worst-case regret of $widetilde O(N5/2sqrtT)$, where $N$ denotes the number of products and $T$ denotes the number of time periods.
- Score: 16.37657820732206
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In addition to maximizing the total revenue, decision-makers in lots of
industries would like to guarantee balanced consumption across different
resources. For instance, in the retailing industry, ensuring a balanced
consumption of resources from different suppliers enhances fairness and helps
main a healthy channel relationship; in the cloud computing industry,
resource-consumption balance helps increase customer satisfaction and reduce
operational costs. Motivated by these practical needs, this paper studies the
price-based network revenue management (NRM) problem with both demand learning
and fair resource-consumption balancing. We introduce the regularized revenue,
i.e., the total revenue with a balancing regularization, as our objective to
incorporate fair resource-consumption balancing into the revenue maximization
goal. We propose a primal-dual-type online policy with the
Upper-Confidence-Bound (UCB) demand learning method to maximize the regularized
revenue. We adopt several innovative techniques to make our algorithm a unified
and computationally efficient framework for the continuous price set and a wide
class of balancing regularizers. Our algorithm achieves a worst-case regret of
$\widetilde O(N^{5/2}\sqrt{T})$, where $N$ denotes the number of products and
$T$ denotes the number of time periods. Numerical experiments in a few NRM
examples demonstrate the effectiveness of our algorithm in simultaneously
achieving revenue maximization and fair resource-consumption balancing
Related papers
- Reinforcement Learning-Based Adaptive Load Balancing for Dynamic Cloud Environments [0.0]
We propose a novel adaptive load balancing framework using Reinforcement Learning (RL) to address these challenges.
Our framework is designed to dynamically reallocate tasks to minimize latency and ensure balanced resource usage across servers.
Experimental results show that the proposed RL-based load balancer outperforms traditional algorithms in terms of response time, resource utilization, and adaptability to changing workloads.
arXiv Detail & Related papers (2024-09-07T19:40:48Z) - Overcoming Recency Bias of Normalization Statistics in Continual
Learning: Balance and Adaptation [67.77048565738728]
Continual learning involves learning a sequence of tasks and balancing their knowledge appropriately.
We propose Adaptive Balance of BN (AdaB$2$N), which incorporates appropriately a Bayesian-based strategy to adapt task-wise contributions.
Our approach achieves significant performance gains across a wide range of benchmarks.
arXiv Detail & Related papers (2023-10-13T04:50:40Z) - Online Learning for Incentive-Based Demand Response [0.0]
We consider the problem of learning online to manage Demand Response (DR) resources.
We propose an online learning scheme that employs least-squares for estimation with a perturbation to the reward price.
arXiv Detail & Related papers (2023-03-27T22:08:05Z) - A State-Augmented Approach for Learning Optimal Resource Management
Decisions in Wireless Networks [58.720142291102135]
We consider a radio resource management (RRM) problem in a multi-user wireless network.
The goal is to optimize a network-wide utility function subject to constraints on the ergodic average performance of users.
We propose a state-augmented parameterization for the RRM policy, where alongside the instantaneous network states, the RRM policy takes as input the set of dual variables corresponding to the constraints.
arXiv Detail & Related papers (2022-10-28T21:24:13Z) - Approaching sales forecasting using recurrent neural networks and
transformers [57.43518732385863]
We develop three alternatives to tackle the problem of forecasting the customer sales at day/store/item level using deep learning techniques.
Our empirical results show how good performance can be achieved by using a simple sequence to sequence architecture with minimal data preprocessing effort.
The proposed solution achieves a RMSLE of around 0.54, which is competitive with other more specific solutions to the problem proposed in the Kaggle competition.
arXiv Detail & Related papers (2022-04-16T12:03:52Z) - Towards Revenue Maximization with Popular and Profitable Products [69.21810902381009]
A common goal for companies marketing is to maximize the return revenue/profit by utilizing the various effective marketing strategies.
Finding credible and reliable information on products' profitability is difficult since most products tends to peak at certain times.
This paper proposes a general profit-oriented framework to address the problem of revenue based on economic behavior, and conducting the 0n-shelf Popular and most Profitable Products (OPPPs) for the targeted marketing.
arXiv Detail & Related papers (2022-02-26T02:07:25Z) - PROMPT: Learning Dynamic Resource Allocation Policies for Network
Applications [16.812611987082082]
We propose PROMPT, a novel resource allocation framework using proactive prediction to guide a reinforcement learning controller.
We show that PROMPT incurs 4.2x fewer violations, reduces severity of policy violations by 12.7x, improves best-effort workload performance, and improves overall power efficiency over prior work.
arXiv Detail & Related papers (2022-01-19T23:34:34Z) - Computation Resource Allocation Solution in Recommender Systems [19.456109814747048]
We propose a computation resource allocation solution (CRAS) that maximizes the business goal with limited computation resources and response time.
The effectiveness of our method is verified by extensive experiments based on the real dataset from Taobao.com.
arXiv Detail & Related papers (2021-03-03T08:41:43Z) - Coordinated Online Learning for Multi-Agent Systems with Coupled
Constraints and Perturbed Utility Observations [91.02019381927236]
We introduce a novel method to steer the agents toward a stable population state, fulfilling the given resource constraints.
The proposed method is a decentralized resource pricing method based on the resource loads resulting from the augmentation of the game's Lagrangian.
arXiv Detail & Related papers (2020-10-21T10:11:17Z) - Resource Allocation via Model-Free Deep Learning in Free Space Optical
Communications [119.81868223344173]
The paper investigates the general problem of resource allocation for mitigating channel fading effects in Free Space Optical (FSO) communications.
Under this framework, we propose two algorithms that solve FSO resource allocation problems.
arXiv Detail & Related papers (2020-07-27T17:38:51Z) - Regularized Online Allocation Problems: Fairness and Beyond [7.433931244705934]
We introduce the emphregularized online allocation problem, a variant that includes a non-linear regularizer acting on the total resource consumption.
In this problem, requests repeatedly arrive over time and, for each request, a decision maker needs to take an action that generates a reward and consumes resources.
The objective is to simultaneously maximize additively separable rewards and the value of a non-separable regularizer subject to the resource constraints.
arXiv Detail & Related papers (2020-07-01T14:24:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.