Graph neural networks-based Scheduler for Production planning problems
using Reinforcement Learning
- URL: http://arxiv.org/abs/2009.03836v2
- Date: Tue, 16 May 2023 10:31:58 GMT
- Title: Graph neural networks-based Scheduler for Production planning problems
using Reinforcement Learning
- Authors: Mohammed Sharafath Abdul Hameed, Andreas Schwung
- Abstract summary: We present a novel framework - GraSP-RL, GRAph neural network-based Scheduler for Production planning problems using Reinforcement Learning.
It represents JSSP as a graph and trains the RL agent using features extracted using a graph neural network (GNN)
While the graph is itself in the non-euclidean space, the features extracted using the GNNs provide a rich encoding of the current production state in the euclidean space, which is then used by the RL agent to select the next job.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Reinforcement learning (RL) is increasingly adopted in job shop scheduling
problems (JSSP). But RL for JSSP is usually done using a vectorized
representation of machine features as the state space. It has three major
problems: (1) the relationship between the machine units and the job sequence
is not fully captured, (2) exponential increase in the size of the state space
with increasing machines/jobs, and (3) the generalization of the agent to
unseen scenarios. We present a novel framework - GraSP-RL, GRAph neural
network-based Scheduler for Production planning problems using Reinforcement
Learning. It represents JSSP as a graph and trains the RL agent using features
extracted using a graph neural network (GNN). While the graph is itself in the
non-euclidean space, the features extracted using the GNNs provide a rich
encoding of the current production state in the euclidean space, which is then
used by the RL agent to select the next job. Further, we cast the scheduling
problem as a decentralized optimization problem in which the learning agent is
assigned to all the production units and the agent learns asynchronously from
the data collected on all the production units. The GraSP-RL is then applied to
a complex injection molding production environment with 30 jobs and 4 machines.
The task is to minimize the makespan of the production plan. The schedule
planned by GraSP-RL is then compared and analyzed with a priority dispatch rule
algorithm like first-in-first-out (FIFO) and metaheuristics like tabu search
(TS) and genetic algorithm (GA). The proposed GraSP-RL outperforms the FIFO,
TS, and GA for the trained task of planning 30 jobs in JSSP. We further test
the generalization capability of the trained agent on two different problem
classes: Open shop system (OSS) and Reactive JSSP (RJSSP) where our method
produces results better than FIFO and comparable results to TS and GA.
Related papers
- StreamRL: Scalable, Heterogeneous, and Elastic RL for LLMs with Disaggregated Stream Generation [55.75008325187133]
Reinforcement learning (RL) has become the core post-training technique for large language models (LLMs)
StreamRL is designed with disaggregation from first principles to address two types of performance bottlenecks.
Experiments show that StreamRL improves throughput by up to 2.66x compared to existing state-of-the-art systems.
arXiv Detail & Related papers (2025-04-22T14:19:06Z) - FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - Learning to Solve Job Shop Scheduling under Uncertainty [1.3002317221601185]
Job-Shop Scheduling Problem (JSSP) is a optimization problem where tasks need to be scheduled on machines in order to minimize criteria such as makespan or delay.
This paper introduces a new approach that leverages Deep Reinforcement Learning (DRL) techniques to search for robust solutions.
arXiv Detail & Related papers (2024-03-04T08:38:55Z) - OTOv3: Automatic Architecture-Agnostic Neural Network Training and
Compression from Structured Pruning to Erasing Operators [57.145175475579315]
This topic spans various techniques, from structured pruning to neural architecture search, encompassing both pruning and erasing operators perspectives.
We introduce the third-generation Only-Train-Once (OTOv3), which first automatically trains and compresses a general DNN through pruning and erasing operations.
Our empirical results demonstrate the efficacy of OTOv3 across various benchmarks in structured pruning and neural architecture search.
arXiv Detail & Related papers (2023-12-15T00:22:55Z) - Learning Regularized Graphon Mean-Field Games with Unknown Graphons [155.38727464526923]
We design reinforcement learning algorithms for Graphon Mean-Field Games (GMFGs)
We aim to learn the Nash Equilibrium (NE) of the regularized GMFGs when the graphons are unknown.
These algorithms are the first specifically designed for learning graphons from sampled agents.
arXiv Detail & Related papers (2023-10-26T16:19:24Z) - Deep reinforcement learning for machine scheduling: Methodology, the
state-of-the-art, and future directions [2.4541568670428915]
Machine scheduling aims to optimize job assignments to machines while adhering to manufacturing rules and job specifications.
Deep Reinforcement Learning (DRL), a key component of artificial general intelligence, has shown promise in various domains like gaming and robotics.
This paper offers a comprehensive review and comparison of DRL-based approaches, highlighting their methodology, applications, advantages, and limitations.
arXiv Detail & Related papers (2023-10-04T22:45:09Z) - HARL: Hierarchical Adaptive Reinforcement Learning Based Auto Scheduler
for Neural Networks [51.71682428015139]
We propose HARL, a reinforcement learning-based auto-scheduler for efficient tensor program exploration.
HarL improves the tensor operator performance by 22% and the search speed by 4.3x compared to the state-of-the-art auto-scheduler.
Inference performance and search speed are also significantly improved on end-to-end neural networks.
arXiv Detail & Related papers (2022-11-21T04:15:27Z) - Learning to schedule job-shop problems: Representation and policy
learning using graph neural network and reinforcement learning [9.379652654427959]
We propose a framework to learn to schedule a job-shop problem (JSSP) using a graph neural network (GNN) and reinforcement learning (RL)
We empirically demonstrate that the GNN scheduler, due to its superb generalization capability, outperforms practically favored rules and RL-based schedulers on various benchmark JSSP.
arXiv Detail & Related papers (2021-06-02T11:40:22Z) - Towards Standardizing Reinforcement Learning Approaches for Stochastic
Production Scheduling [77.34726150561087]
reinforcement learning can be used to solve scheduling problems.
Existing studies rely on (sometimes) complex simulations for which the code is unavailable.
There is a vast array of RL designs to choose from.
standardization of model descriptions - both production setup and RL design - and validation scheme are a prerequisite.
arXiv Detail & Related papers (2021-04-16T16:07:10Z) - RL-Scope: Cross-Stack Profiling for Deep Reinforcement Learning
Workloads [4.575381867242508]
We propose RL-Scope, a cross-stack profiler that scopes low-level CPU/GPU resource usage to high-level algorithmic operations.
We demonstrate RL-Scope's utility through in-depth case studies.
arXiv Detail & Related papers (2021-02-08T15:42:48Z) - L$^2$-GCN: Layer-Wise and Learned Efficient Training of Graph
Convolutional Networks [118.37805042816784]
Graph convolution networks (GCN) are increasingly popular in many applications, yet remain notoriously hard to train over large graph datasets.
We propose a novel efficient layer-wise training framework for GCN (L-GCN), that disentangles feature aggregation and feature transformation during training.
Experiments show that L-GCN is faster than state-of-the-arts by at least an order of magnitude, with a consistent of memory usage not dependent on dataset size.
arXiv Detail & Related papers (2020-03-30T16:37:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.