Deep Reinforcement Learning for Resource Allocation in Business
Processes
- URL: http://arxiv.org/abs/2104.00541v1
- Date: Mon, 29 Mar 2021 11:20:25 GMT
- Title: Deep Reinforcement Learning for Resource Allocation in Business
Processes
- Authors: Kamil \.Zbikowski, Micha{\l} Ostapowicz, Piotr Gawrysiak
- Abstract summary: We propose a novel representation that allows modeling of a multi-process environment with different process-based rewards.
We then use double deep reinforcement learning to look for optimal resource allocation policy.
Deep reinforcement learning based resource allocation achieved significantly better results than two commonly used techniques.
- Score: 3.0938904602244355
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Assigning resources in business processes execution is a repetitive task that
can be effectively automated. However, different automation methods may give
varying results that may not be optimal. Proper resource allocation is crucial
as it may lead to significant cost reductions or increased effectiveness that
results in increased revenues.
In this work, we first propose a novel representation that allows modeling of
a multi-process environment with different process-based rewards. These
processes can share resources that differ in their eligibility. Then, we use
double deep reinforcement learning to look for optimal resource allocation
policy. We compare those results with two popular strategies that are widely
used in the industry. Learning optimal policy through reinforcement learning
requires frequent interactions with the environment, so we also designed and
developed a simulation engine that can mimic real-world processes.
The results obtained are promising. Deep reinforcement learning based
resource allocation achieved significantly better results compared to two
commonly used techniques.
Related papers
- From Novice to Expert: LLM Agent Policy Optimization via Step-wise Reinforcement Learning [62.54484062185869]
We introduce StepAgent, which utilizes step-wise reward to optimize the agent's reinforcement learning process.
We propose implicit-reward and inverse reinforcement learning techniques to facilitate agent reflection and policy adjustment.
arXiv Detail & Related papers (2024-11-06T10:35:11Z) - Sample Efficient Reinforcement Learning by Automatically Learning to
Compose Subtasks [3.1594865504808944]
We propose an RL algorithm that automatically structure the reward function for sample efficiency, given a set of labels that signify subtasks.
We evaluate our algorithm in a variety of sparse-reward environments.
arXiv Detail & Related papers (2024-01-25T15:06:40Z) - RLIF: Interactive Imitation Learning as Reinforcement Learning [56.997263135104504]
We show how off-policy reinforcement learning can enable improved performance under assumptions that are similar but potentially even more practical than those of interactive imitation learning.
Our proposed method uses reinforcement learning with user intervention signals themselves as rewards.
This relaxes the assumption that intervening experts in interactive imitation learning should be near-optimal and enables the algorithm to learn behaviors that improve over the potential suboptimal human expert.
arXiv Detail & Related papers (2023-11-21T21:05:21Z) - Reinforcement Learning with Foundation Priors: Let the Embodied Agent Efficiently Learn on Its Own [59.11934130045106]
We propose Reinforcement Learning with Foundation Priors (RLFP) to utilize guidance and feedback from policy, value, and success-reward foundation models.
Within this framework, we introduce the Foundation-guided Actor-Critic (FAC) algorithm, which enables embodied agents to explore more efficiently with automatic reward functions.
Our method achieves remarkable performances in various manipulation tasks on both real robots and in simulation.
arXiv Detail & Related papers (2023-10-04T07:56:42Z) - Learning policies for resource allocation in business processes [0.0]
This paper proposes two learning-based methods for resource allocation in business processes.
The first method leverages Deep Reinforcement Learning (DRL) to learn near-optimal policies by taking action in the business process.
The second method is a score-based value approximation approach, which learns the weights of a set of curated features to prioritize resource assignments.
arXiv Detail & Related papers (2023-04-19T21:05:38Z) - Dynamic Resource Allocation for Metaverse Applications with Deep
Reinforcement Learning [64.75603723249837]
This work proposes a novel framework to dynamically manage and allocate different types of resources for Metaverse applications.
We first propose an effective solution to divide applications into groups, namely MetaInstances, where common functions can be shared among applications.
Then, to capture the real-time, dynamic, and uncertain characteristics of request arrival and application departure processes, we develop a semi-Markov decision process-based framework.
arXiv Detail & Related papers (2023-02-27T00:30:01Z) - Learning to Optimize for Reinforcement Learning [58.01132862590378]
Reinforcement learning (RL) is essentially different from supervised learning, and in practice, these learneds do not work well even in simple RL tasks.
Agent-gradient distribution is non-independent and identically distributed, leading to inefficient meta-training.
We show that, although only trained in toy tasks, our learned can generalize unseen complex tasks in Brax.
arXiv Detail & Related papers (2023-02-03T00:11:02Z) - The Cost of Learning: Efficiency vs. Efficacy of Learning-Based RRM for
6G [10.28841351455586]
Deep Reinforcement Learning (DRL) has become a valuable solution to automatically learn efficient resource management strategies in complex networks.
In many scenarios, the learning task is performed in the Cloud, while experience samples are generated directly by edge nodes or users.
This creates a friction between the need to speed up convergence towards an effective strategy, which requires the allocation of resources to transmit learning samples.
We propose a dynamic balancing strategy between the learning and data planes, which allows the centralized learning agent to quickly converge to an efficient resource allocation strategy.
arXiv Detail & Related papers (2022-11-30T11:26:01Z) - Actively Learning Costly Reward Functions for Reinforcement Learning [56.34005280792013]
We show that it is possible to train agents in complex real-world environments orders of magnitudes faster.
By enabling the application of reinforcement learning methods to new domains, we show that we can find interesting and non-trivial solutions.
arXiv Detail & Related papers (2022-11-23T19:17:20Z) - Modelling resource allocation in uncertain system environment through
deep reinforcement learning [0.0]
Reinforcement Learning has applications in field of mechatronics, robotics, and other resource-constrained control system.
The paper identifies problem of resource allocation in uncertain environment could be effectively solved using Noisy Bagging duelling double deep Q network.
arXiv Detail & Related papers (2021-06-17T13:13:34Z) - Efficient Reinforcement Learning in Resource Allocation Problems Through
Permutation Invariant Multi-task Learning [6.247939901619901]
We show that in certain settings, the available data can be dramatically increased through a form of multi-task learning.
We provide a theoretical performance bound for the gain in sample efficiency under this setting.
This motivates a new approach to multi-task learning, which involves the design of an appropriate neural network architecture and a prioritized task-sampling strategy.
arXiv Detail & Related papers (2021-02-18T14:13:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.