Dynamic Dispatching for Large-Scale Heterogeneous Fleet via Multi-agent
Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2008.10713v1
- Date: Mon, 24 Aug 2020 21:29:56 GMT
- Title: Dynamic Dispatching for Large-Scale Heterogeneous Fleet via Multi-agent
Deep Reinforcement Learning
- Authors: Chi Zhang, Philip Odonkor, Shuai Zheng, Hamed Khorasgani, Susumu
Serita, Chetan Gupta
- Abstract summary: We propose a novel Deep Reinforcement Learning approach to solve the dynamic problem in mining.
We first develop an event-based mining simulator with parameters calibrated in real mines.
Then we propose an experience-sharing Deep Q Network with a novel abstract state/action representation to learn memories from heterogeneous agents.
- Score: 10.835960004409708
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dynamic dispatching is one of the core problems for operation optimization in
traditional industries such as mining, as it is about how to smartly allocate
the right resources to the right place at the right time. Conventionally, the
industry relies on heuristics or even human intuitions which are often
short-sighted and sub-optimal solutions. Leveraging the power of AI and
Internet of Things (IoT), data-driven automation is reshaping this area.
However, facing its own challenges such as large-scale and heterogenous trucks
running in a highly dynamic environment, it can barely adopt methods developed
in other domains (e.g., ride-sharing). In this paper, we propose a novel Deep
Reinforcement Learning approach to solve the dynamic dispatching problem in
mining. We first develop an event-based mining simulator with parameters
calibrated in real mines. Then we propose an experience-sharing Deep Q Network
with a novel abstract state/action representation to learn memories from
heterogeneous agents altogether and realizes learning in a centralized way. We
demonstrate that the proposed methods significantly outperform the most widely
adopted approaches in the industry by $5.56\%$ in terms of productivity. The
proposed approach has great potential in a broader range of industries (e.g.,
manufacturing, logistics) which have a large-scale of heterogenous equipment
working in a highly dynamic environment, as a general framework for dynamic
resource allocation.
Related papers
- Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - Imagine, Initialize, and Explore: An Effective Exploration Method in
Multi-Agent Reinforcement Learning [27.81925751697255]
We propose a novel method for efficient multi-agent exploration in complex scenarios.
We formulate the imagination as a sequence modeling problem, where the states, observations, prompts, actions, and rewards are predicted autoregressively.
By initializing agents at the critical states, IIE significantly increases the likelihood of discovering potentially important underexplored regions.
arXiv Detail & Related papers (2024-02-28T01:45:01Z) - An Interactive Agent Foundation Model [49.77861810045509]
We propose an Interactive Agent Foundation Model that uses a novel multi-task agent training paradigm for training AI agents.
Our training paradigm unifies diverse pre-training strategies, including visual masked auto-encoders, language modeling, and next-action prediction.
We demonstrate the performance of our framework across three separate domains -- Robotics, Gaming AI, and Healthcare.
arXiv Detail & Related papers (2024-02-08T18:58:02Z) - DREAM: Decentralized Reinforcement Learning for Exploration and
Efficient Energy Management in Multi-Robot Systems [14.266876062352424]
Resource-constrained robots often suffer from energy inefficiencies, underutilized computational abilities due to inadequate task allocation, and a lack of robustness in dynamic environments.
This paper introduces DREAM - Decentralized Reinforcement Learning for Exploration and Efficient Energy Management in Multi-Robot Systems.
arXiv Detail & Related papers (2023-09-29T17:43:41Z) - Locality Matters: A Scalable Value Decomposition Approach for
Cooperative Multi-Agent Reinforcement Learning [52.7873574425376]
Cooperative multi-agent reinforcement learning (MARL) faces significant scalability issues due to state and action spaces that are exponentially large in the number of agents.
We propose a novel, value-based multi-agent algorithm called LOMAQ, which incorporates local rewards in the Training Decentralized Execution paradigm.
arXiv Detail & Related papers (2021-09-22T10:08:15Z) - Efficient Model-Based Multi-Agent Mean-Field Reinforcement Learning [89.31889875864599]
We propose an efficient model-based reinforcement learning algorithm for learning in multi-agent systems.
Our main theoretical contributions are the first general regret bounds for model-based reinforcement learning for MFC.
We provide a practical parametrization of the core optimization problem.
arXiv Detail & Related papers (2021-07-08T18:01:02Z) - Graph Neural Networks for Decentralized Multi-Robot Submodular Action
Selection [101.38634057635373]
We focus on applications where robots are required to jointly select actions to maximize team submodular objectives.
We propose a general-purpose learning architecture towards submodular at scale, with decentralized communications.
We demonstrate the performance of our GNN-based learning approach in a scenario of active target coverage with large networks of robots.
arXiv Detail & Related papers (2021-05-18T15:32:07Z) - Challenges of Applying Deep Reinforcement Learning in Dynamic
Dispatching [14.412373060545491]
Dynamic dispatching is one of the core problems for operations optimization in the mining industry.
Deep reinforcement learning should be a natural fit to solve this problem.
In this paper, we review the main challenges in using deep RL to address the dynamic dispatching problem.
arXiv Detail & Related papers (2020-11-09T22:26:45Z) - Dif-MAML: Decentralized Multi-Agent Meta-Learning [54.39661018886268]
We propose a cooperative multi-agent meta-learning algorithm, referred to as MAML or Dif-MAML.
We show that the proposed strategy allows a collection of agents to attain agreement at a linear rate and to converge to a stationary point of the aggregate MAML.
Simulation results illustrate the theoretical findings and the superior performance relative to the traditional non-cooperative setting.
arXiv Detail & Related papers (2020-10-06T16:51:09Z) - MAPPER: Multi-Agent Path Planning with Evolutionary Reinforcement
Learning in Mixed Dynamic Environments [30.407700996710023]
This paper proposes a decentralized partially observable multi-agent path planning with evolutionary reinforcement learning (MAPPER) method.
We decompose the long-range navigation task into many easier sub-tasks under the guidance of a global planner.
Our approach models dynamic obstacles' behavior with an image-based representation and trains a policy in mixed dynamic environments without homogeneity assumption.
arXiv Detail & Related papers (2020-07-30T20:14:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.