ScheduleNet: Learn to solve multi-agent scheduling problems with
reinforcement learning
- URL: http://arxiv.org/abs/2106.03051v1
- Date: Sun, 6 Jun 2021 07:08:58 GMT
- Title: ScheduleNet: Learn to solve multi-agent scheduling problems with
reinforcement learning
- Authors: Junyoung Park, Sanjar Bakhtiyar, Jinkyoo Park
- Abstract summary: We learn ScheduleNet, a decentralized decision-making policy that can effectively coordinate multiple agents to complete tasks.
We validate the effectiveness of ScheduleNet as a general learning-based scheduler for solving various types of multi-agent scheduling tasks.
- Score: 10.16257074782054
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose ScheduleNet, a RL-based real-time scheduler, that can solve
various types of multi-agent scheduling problems. We formulate these problems
as a semi-MDP with episodic reward (makespan) and learn ScheduleNet, a
decentralized decision-making policy that can effectively coordinate multiple
agents to complete tasks. The decision making procedure of ScheduleNet
includes: (1) representing the state of a scheduling problem with the
agent-task graph, (2) extracting node embeddings for agent and tasks nodes, the
important relational information among agents and tasks, by employing the
type-aware graph attention (TGA), and (3) computing the assignment probability
with the computed node embeddings. We validate the effectiveness of ScheduleNet
as a general learning-based scheduler for solving various types of multi-agent
scheduling tasks, including multiple salesman traveling problem (mTSP) and job
shop scheduling problem (JSP).
Related papers
- Agent-Oriented Planning in Multi-Agent Systems [54.429028104022066]
We propose a novel framework for agent-oriented planning in multi-agent systems, leveraging a fast task decomposition and allocation process.
We integrate a feedback loop into the proposed framework to further enhance the effectiveness and robustness of such a problem-solving process.
arXiv Detail & Related papers (2024-10-03T04:07:51Z) - Reinforcement Learning with Success Induced Task Prioritization [68.8204255655161]
We introduce Success Induced Task Prioritization (SITP), a framework for automatic curriculum learning.
The algorithm selects the order of tasks that provide the fastest learning for agents.
We demonstrate that SITP matches or surpasses the results of other curriculum design methods.
arXiv Detail & Related papers (2022-12-30T12:32:43Z) - Meta-learning with an Adaptive Task Scheduler [93.63502984214918]
Existing meta-learning algorithms randomly sample meta-training tasks with a uniform probability.
It is likely that tasks are detrimental with noise or imbalanced given a limited number of meta-training tasks.
We propose an adaptive task scheduler (ATS) for the meta-training process.
arXiv Detail & Related papers (2021-10-26T22:16:35Z) - An Evolutionary Algorithm for Task Scheduling in Crowdsourced Software
Development [10.373891804761376]
This paper proposes an evolutionary algorithm-based task scheduling method for crowdsourced software development.
Experimental results on 4 projects demonstrate that the proposed method has the potential to reduce project duration by a factor of 33-78%.
arXiv Detail & Related papers (2021-07-05T18:07:26Z) - Better than the Best: Gradient-based Improper Reinforcement Learning for
Network Scheduling [60.48359567964899]
We consider the problem of scheduling in constrained queueing networks with a view to minimizing packet delay.
We use a policy gradient based reinforcement learning algorithm that produces a scheduler that performs better than the available atomic policies.
arXiv Detail & Related papers (2021-05-01T10:18:34Z) - Scheduling Plans of Tasks [0.0]
We present a algorithm for solving the problem of scheduling plans of tasks.
The proposed algorithm searches for a feasible schedule that maximize the number of plans scheduled.
arXiv Detail & Related papers (2021-02-06T10:14:54Z) - Scalable Anytime Planning for Multi-Agent MDPs [37.69939216970677]
We present a scalable tree search planning algorithm for large multi-agent sequential decision problems that require dynamic collaboration.
Our algorithm comprises three elements: online planning with Monte Carlo Tree Search (MCTS), factored representations of local agent interactions with coordination graphs, and the iterative Max-Plus method for joint action selection.
arXiv Detail & Related papers (2021-01-12T22:50:17Z) - A Machine Learning Approach for Task and Resource Allocation in Mobile
Edge Computing Based Networks [108.57859531628264]
A joint task, spectrum, and transmit power allocation problem is investigated for a wireless network.
The proposed algorithm can reduce the number of iterations needed for convergence and the maximal delay among all users by up to 18% and 11.1% compared to the standard Q-learning algorithm.
arXiv Detail & Related papers (2020-07-20T13:46:42Z) - Dynamic Multi-Robot Task Allocation under Uncertainty and Temporal
Constraints [52.58352707495122]
We present a multi-robot allocation algorithm that decouples the key computational challenges of sequential decision-making under uncertainty and multi-agent coordination.
We validate our results over a wide range of simulations on two distinct domains: multi-arm conveyor belt pick-and-place and multi-drone delivery dispatch in a city.
arXiv Detail & Related papers (2020-05-27T01:10:41Z) - A Novel Multi-Agent System for Complex Scheduling Problems [2.294014185517203]
This paper is the conception and implementation of a multi-agent system that is applicable in various problem domains.
We simulate a NP-hard scheduling problem to demonstrate the validity of our approach.
This paper highlights the advantages of the agent-based approach, like the reduction in layout complexity, improved control of complicated systems, and extendability.
arXiv Detail & Related papers (2020-04-20T14:04:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.