Inference-Based Deterministic Messaging For Multi-Agent Communication
- URL: http://arxiv.org/abs/2103.02150v1
- Date: Wed, 3 Mar 2021 03:09:22 GMT
- Title: Inference-Based Deterministic Messaging For Multi-Agent Communication
- Authors: Varun Bhatt, Michael Buro
- Abstract summary: We study learning in matrix-based signaling games to show that decentralized methods can converge to a suboptimal policy.
We then propose a modification to the messaging policy, in which the sender deterministically chooses the best message that helps the receiver to infer the sender's observation.
- Score: 1.8275108630751844
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Communication is essential for coordination among humans and animals.
Therefore, with the introduction of intelligent agents into the world,
agent-to-agent and agent-to-human communication becomes necessary. In this
paper, we first study learning in matrix-based signaling games to empirically
show that decentralized methods can converge to a suboptimal policy. We then
propose a modification to the messaging policy, in which the sender
deterministically chooses the best message that helps the receiver to infer the
sender's observation. Using this modification, we see, empirically, that the
agents converge to the optimal policy in nearly all the runs. We then apply
this method to a partially observable gridworld environment which requires
cooperation between two agents and show that, with appropriate approximation
methods, the proposed sender modification can enhance existing decentralized
training methods for more complex domains as well.
Related papers
- Communication Learning in Multi-Agent Systems from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
We introduce a temporal gating mechanism for each agent, enabling dynamic decisions on whether to receive shared information at a given time.
arXiv Detail & Related papers (2024-11-01T05:56:51Z) - DCMAC: Demand-aware Customized Multi-Agent Communication via Upper Bound Training [9.068971933560416]
We propose a Demand-aware Customized Multi-Agent Communication protocol, which use an upper bound training to obtain the ideal policy.
Experimental results reveal that DCMAC significantly outperforms the baseline algorithms in both unconstrained and communication constrained scenarios.
arXiv Detail & Related papers (2024-09-11T09:23:27Z) - Learning Multi-Agent Communication from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
Our proposed approach, CommFormer, efficiently optimize the communication graph and concurrently refines architectural parameters through gradient descent in an end-to-end manner.
arXiv Detail & Related papers (2024-05-14T12:40:25Z) - Pragmatic Communication in Multi-Agent Collaborative Perception [80.14322755297788]
Collaborative perception results in a trade-off between perception ability and communication costs.
We propose PragComm, a multi-agent collaborative perception system with two key components.
PragComm consistently outperforms previous methods with more than 32.7K times lower communication volume.
arXiv Detail & Related papers (2024-01-23T11:58:08Z) - Enhancing Multi-Agent Coordination through Common Operating Picture
Integration [14.927199437011044]
We present an approach to multi-agent coordination, where each agent is equipped with the capability to integrate its history of observations, actions and messages received into a Common Operating Picture (COP)
Our results demonstrate the efficacy of COP integration, and show that COP-based training leads to robust policies compared to state-of-the-art Multi-Agent Reinforcement Learning (MARL) methods when faced with out-of-distribution initial states.
arXiv Detail & Related papers (2023-11-08T15:08:55Z) - Research on Multi-Agent Communication and Collaborative Decision-Making
Based on Deep Reinforcement Learning [0.0]
This thesis studies the cooperative decision-making of multi-agent based on the Multi-Agent Proximal Policy Optimization algorithm.
Different agents can alleviate the non-stationarity caused by local observations through information exchange between agents.
The experimental results show that the improvement has achieved certain effects, which can better alleviate the non-stationarity of the multi-agent environment.
arXiv Detail & Related papers (2023-05-23T14:20:14Z) - Coordinating Policies Among Multiple Agents via an Intelligent
Communication Channel [81.39444892747512]
In Multi-Agent Reinforcement Learning (MARL), specialized channels are often introduced that allow agents to communicate directly with one another.
We propose an alternative approach whereby agents communicate through an intelligent facilitator that learns to sift through and interpret signals provided by all agents to improve the agents' collective performance.
arXiv Detail & Related papers (2022-05-21T14:11:33Z) - Depthwise Convolution for Multi-Agent Communication with Enhanced
Mean-Field Approximation [9.854975702211165]
We propose a new method based on local communication learning to tackle the multi-agent RL (MARL) challenge.
First, we design a new communication protocol that exploits the ability of depthwise convolution to efficiently extract local relations.
Second, we introduce the mean-field approximation into our method to reduce the scale of agent interactions.
arXiv Detail & Related papers (2022-03-06T07:42:43Z) - Explaining Reinforcement Learning Policies through Counterfactual
Trajectories [147.7246109100945]
A human developer must validate that an RL agent will perform well at test-time.
Our method conveys how the agent performs under distribution shifts by showing the agent's behavior across a wider trajectory distribution.
In a user study, we demonstrate that our method enables users to score better than baseline methods on one of two agent validation tasks.
arXiv Detail & Related papers (2022-01-29T00:52:37Z) - Distributed Adaptive Learning Under Communication Constraints [54.22472738551687]
This work examines adaptive distributed learning strategies designed to operate under communication constraints.
We consider a network of agents that must solve an online optimization problem from continual observation of streaming data.
arXiv Detail & Related papers (2021-12-03T19:23:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.