Minimizing Communication while Maximizing Performance in Multi-Agent
Reinforcement Learning
- URL: http://arxiv.org/abs/2106.08482v1
- Date: Tue, 15 Jun 2021 23:13:51 GMT
- Title: Minimizing Communication while Maximizing Performance in Multi-Agent
Reinforcement Learning
- Authors: Varun Kumar Vijay and Hassam Sheikh and Somdeb Majumdar and Mariano
Phielipp
- Abstract summary: Inter-agent communication can significantly increase performance in multi-agent tasks that require co-ordination.
In real-world applications, where communication may be limited by system constraints like bandwidth, power and network capacity, one might need to reduce the number of messages that are sent.
We show that we can reduce communication by 75% with no loss of performance.
- Score: 5.612141846711729
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inter-agent communication can significantly increase performance in
multi-agent tasks that require co-ordination to achieve a shared goal. Prior
work has shown that it is possible to learn inter-agent communication protocols
using multi-agent reinforcement learning and message-passing network
architectures. However, these models use an unconstrained broadcast
communication model, in which an agent communicates with all other agents at
every step, even when the task does not require it. In real-world applications,
where communication may be limited by system constraints like bandwidth, power
and network capacity, one might need to reduce the number of messages that are
sent. In this work, we explore a simple method of minimizing communication
while maximizing performance in multi-task learning: simultaneously optimizing
a task-specific objective and a communication penalty. We show that the
objectives can be optimized using Reinforce and the Gumbel-Softmax
reparameterization. We introduce two techniques to stabilize training: 50%
training and message forwarding. Training with the communication penalty on
only 50% of the episodes prevents our models from turning off their outgoing
messages. Second, repeating messages received previously helps models retain
information, and further improves performance. With these techniques, we show
that we can reduce communication by 75% with no loss of performance.
Related papers
- Communication Learning in Multi-Agent Systems from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
We introduce a temporal gating mechanism for each agent, enabling dynamic decisions on whether to receive shared information at a given time.
arXiv Detail & Related papers (2024-11-01T05:56:51Z) - Learning Multi-Agent Communication from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
Our proposed approach, CommFormer, efficiently optimize the communication graph and concurrently refines architectural parameters through gradient descent in an end-to-end manner.
arXiv Detail & Related papers (2024-05-14T12:40:25Z) - Context-aware Communication for Multi-agent Reinforcement Learning [6.109127175562235]
We develop a context-aware communication scheme for multi-agent reinforcement learning (MARL)
In the first stage, agents exchange coarse representations in a broadcast fashion, providing context for the second stage.
Following this, agents utilize attention mechanisms in the second stage to selectively generate messages personalized for the receivers.
To evaluate the effectiveness of CACOM, we integrate it with both actor-critic and value-based MARL algorithms.
arXiv Detail & Related papers (2023-12-25T03:33:08Z) - Multi-Receiver Task-Oriented Communications via Multi-Task Deep Learning [49.83882366499547]
This paper studies task-oriented, otherwise known as goal-oriented, communications in a setting where a transmitter communicates with multiple receivers.
A multi-task deep learning approach is presented for joint optimization of completing multiple tasks and communicating with multiple receivers.
arXiv Detail & Related papers (2023-08-14T01:34:34Z) - Towards True Lossless Sparse Communication in Multi-Agent Systems [1.911678487931003]
Communication enables agents to cooperate to achieve their goals.
Recent work in learning sparse individualized communication suffers from high variance during training.
We use the information bottleneck to reframe sparsity as a representation learning problem.
arXiv Detail & Related papers (2022-11-30T20:43:34Z) - Over-communicate no more: Situated RL agents learn concise communication
protocols [78.28898217947467]
It is unclear how to design artificial agents that can learn to effectively and efficiently communicate with each other.
Much research on communication emergence uses reinforcement learning (RL)
We explore situated communication in a multi-step task, where the acting agent has to forgo an environmental action to communicate.
We find that while all tested pressures can disincentivise over-communication, situated communication does it most effectively and, unlike the cost on effort, does not negatively impact emergence.
arXiv Detail & Related papers (2022-11-02T21:08:14Z) - Learning Practical Communication Strategies in Cooperative Multi-Agent
Reinforcement Learning [5.539117319607963]
Communication in realistic wireless networks can be highly unreliable due to network conditions varying with agents' mobility.
We propose a framework to learn practical communication strategies by addressing three fundamental questions.
We show significant improvements in game performance, convergence speed and communication efficiency compared with state-of-the-art.
arXiv Detail & Related papers (2022-09-02T22:18:43Z) - FCMNet: Full Communication Memory Net for Team-Level Cooperation in
Multi-Agent Systems [15.631744703803806]
We introduce FCMNet, a reinforcement learning based approach that allows agents to simultaneously learn an effective multi-hop communications protocol.
Using a simple multi-hop topology, we endow each agent with the ability to receive information sequentially encoded by every other agent at each time step.
FCMNet outperforms state-of-the-art communication-based reinforcement learning methods in all StarCraft II micromanagement tasks.
arXiv Detail & Related papers (2022-01-28T09:12:01Z) - Multi-agent Communication with Graph Information Bottleneck under
Limited Bandwidth (a position paper) [92.11330289225981]
In many real-world scenarios, communication can be expensive and the bandwidth of the multi-agent system is subject to certain constraints.
Redundant messages who occupy the communication resources can block the transmission of informative messages and thus jeopardize the performance.
We propose a novel multi-agent communication module, CommGIB, which effectively compresses the structure information and node information in the communication graph to deal with bandwidth-constrained settings.
arXiv Detail & Related papers (2021-12-20T07:53:44Z) - Learning Individually Inferred Communication for Multi-Agent Cooperation [37.56115000150748]
We propose Individually Inferred Communication (I2C) to enable agents to learn a prior for agent-agent communication.
The prior knowledge is learned via causal inference and realized by a feed-forward neural network.
I2C can not only reduce communication overhead but also improve the performance in a variety of multi-agent cooperative scenarios.
arXiv Detail & Related papers (2020-06-11T14:07:57Z) - Learning Structured Communication for Multi-agent Reinforcement Learning [104.64584573546524]
This work explores the large-scale multi-agent communication mechanism under a multi-agent reinforcement learning (MARL) setting.
We propose a novel framework termed as Learning Structured Communication (LSC) by using a more flexible and efficient communication topology.
arXiv Detail & Related papers (2020-02-11T07:19:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.