Decentralized scheduling through an adaptive, trading-based multi-agent
system
- URL: http://arxiv.org/abs/2207.11172v1
- Date: Tue, 5 Jul 2022 13:50:18 GMT
- Title: Decentralized scheduling through an adaptive, trading-based multi-agent
system
- Authors: Michael K\"olle, Lennart Rietdorf, Kyrill Schmid
- Abstract summary: In multi-agent reinforcement learning systems, the actions of one agent can have a negative impact on the rewards of other agents.
This work applies a trading approach to a simulated scheduling environment, where the agents are responsible for the assignment of incoming jobs to compute cores.
The agents can trade the usage right of computational cores to process high-priority, high-reward jobs faster than low-priority, low-reward jobs.
- Score: 1.7403133838762448
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In multi-agent reinforcement learning systems, the actions of one agent can
have a negative impact on the rewards of other agents. One way to combat this
problem is to let agents trade their rewards amongst each other. Motivated by
this, this work applies a trading approach to a simulated scheduling
environment, where the agents are responsible for the assignment of incoming
jobs to compute cores. In this environment, reinforcement learning agents learn
to trade successfully. The agents can trade the usage right of computational
cores to process high-priority, high-reward jobs faster than low-priority,
low-reward jobs. However, due to combinatorial effects, the action and
observation spaces of a simple reinforcement learning agent in this environment
scale exponentially with key parameters of the problem size. However, the
exponential scaling behavior can be transformed into a linear one if the agent
is split into several independent sub-units. We further improve this
distributed architecture using agent-internal parameter sharing. Moreover, it
can be extended to set the exchange prices autonomously. We show that in our
scheduling environment, the advantages of a distributed agent architecture
clearly outweigh more aggregated approaches. We demonstrate that the
distributed agent architecture becomes even more performant using
agent-internal parameter sharing. Finally, we investigate how two different
reward functions affect autonomous pricing and the corresponding scheduling.
Related papers
- From Novice to Expert: LLM Agent Policy Optimization via Step-wise Reinforcement Learning [62.54484062185869]
We introduce StepAgent, which utilizes step-wise reward to optimize the agent's reinforcement learning process.
We propose implicit-reward and inverse reinforcement learning techniques to facilitate agent reflection and policy adjustment.
arXiv Detail & Related papers (2024-11-06T10:35:11Z) - Strategic Classification With Externalities [11.36782598786846]
We propose a new variant of the strategic classification problem.
Motivated by real-world applications, our model crucially allows the manipulation of one agent to affect another.
We show that under certain assumptions, the pure Nash Equilibrium of this agent manipulation game is unique and can be efficiently computed.
arXiv Detail & Related papers (2024-10-10T15:28:04Z) - AgentVerse: Facilitating Multi-Agent Collaboration and Exploring
Emergent Behaviors [93.38830440346783]
We propose a multi-agent framework framework that can collaboratively adjust its composition as a greater-than-the-sum-of-its-parts system.
Our experiments demonstrate that framework framework can effectively deploy multi-agent groups that outperform a single agent.
In view of these behaviors, we discuss some possible strategies to leverage positive ones and mitigate negative ones for improving the collaborative potential of multi-agent groups.
arXiv Detail & Related papers (2023-08-21T16:47:11Z) - Learning Individual Policies in Large Multi-agent Systems through Local
Variance Minimization [8.140037969280716]
In multi-agent systems with large number of agents, contribution of each agent to value of other agents is minimal.
We provide a novel Multi-Agent Reinforcement Learning (MARL) mechanism that minimizes variance across values of agents in the same state.
We show that our approach reduces the variance in revenues earned by taxi drivers, while still providing higher joint revenues than leading approaches.
arXiv Detail & Related papers (2022-12-27T06:59:00Z) - Explaining Reinforcement Learning Policies through Counterfactual
Trajectories [147.7246109100945]
A human developer must validate that an RL agent will perform well at test-time.
Our method conveys how the agent performs under distribution shifts by showing the agent's behavior across a wider trajectory distribution.
In a user study, we demonstrate that our method enables users to score better than baseline methods on one of two agent validation tasks.
arXiv Detail & Related papers (2022-01-29T00:52:37Z) - Distributed Adaptive Learning Under Communication Constraints [54.22472738551687]
This work examines adaptive distributed learning strategies designed to operate under communication constraints.
We consider a network of agents that must solve an online optimization problem from continual observation of streaming data.
arXiv Detail & Related papers (2021-12-03T19:23:48Z) - Interactive Inverse Reinforcement Learning for Cooperative Games [7.257751371276486]
We study the problem of designing AI agents that can learn to cooperate effectively with a potentially suboptimal partner.
This problem is modeled as a cooperative episodic two-agent Markov decision process.
We show that when the learning agent's policies have a significant effect on the transition function, the reward function can be learned efficiently.
arXiv Detail & Related papers (2021-11-08T18:24:52Z) - BGC: Multi-Agent Group Belief with Graph Clustering [1.9949730506194252]
We propose a semi-communication method to enable agents can exchange information without communication.
Inspired by the neighborhood cognitive consistency, we propose a group-based module to divide adjacent agents into a small group and minimize in-group agents' beliefs.
Results reveal that the proposed method achieves a significant improvement in the SMAC benchmark.
arXiv Detail & Related papers (2020-08-20T07:07:20Z) - Randomized Entity-wise Factorization for Multi-Agent Reinforcement
Learning [59.62721526353915]
Multi-agent settings in the real world often involve tasks with varying types and quantities of agents and non-agent entities.
Our method aims to leverage these commonalities by asking the question: What is the expected utility of each agent when only considering a randomly selected sub-group of its observed entities?''
arXiv Detail & Related papers (2020-06-07T18:28:41Z) - Counterfactual Multi-Agent Reinforcement Learning with Graph Convolution
Communication [5.5438676149999075]
We consider a fully cooperative multi-agent system where agents cooperate to maximize a system's utility.
We propose that multi-agent systems must have the ability to communicate and understand the inter-plays between agents.
We develop an architecture that allows for communication among agents and tailors the system's reward for each individual agent.
arXiv Detail & Related papers (2020-04-01T14:36:13Z) - Scalable Multi-Agent Inverse Reinforcement Learning via
Actor-Attention-Critic [54.2180984002807]
Multi-agent adversarial inverse reinforcement learning (MA-AIRL) is a recent approach that applies single-agent AIRL to multi-agent problems.
We propose a multi-agent inverse RL algorithm that is more sample-efficient and scalable than previous works.
arXiv Detail & Related papers (2020-02-24T20:30:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.