Cooperative Multi-Agent Learning for Navigation via Structured State
Abstraction
- URL: http://arxiv.org/abs/2306.11336v2
- Date: Mon, 12 Feb 2024 08:06:28 GMT
- Title: Cooperative Multi-Agent Learning for Navigation via Structured State
Abstraction
- Authors: Mohamed K. Abdelaziz, Mohammed S. Elbamby, Sumudu Samarakoon, Mehdi
Bennis
- Abstract summary: This work proposes a novel neural network architecture for learning an adaptive state space abstraction and a communication protocol among agents participating in navigation tasks.
The goal is to come up with an adaptive abstractor that significantly reduces the size of the state space to be explored, without degradation in the policy performance.
- Score: 35.137788512033694
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cooperative multi-agent reinforcement learning (MARL) for navigation enables
agents to cooperate to achieve their navigation goals. Using emergent
communication, agents learn a communication protocol to coordinate and share
information that is needed to achieve their navigation tasks. In emergent
communication, symbols with no pre-specified usage rules are exchanged, in
which the meaning and syntax emerge through training. Learning a navigation
policy along with a communication protocol in a MARL environment is highly
complex due to the huge state space to be explored. To cope with this
complexity, this work proposes a novel neural network architecture, for jointly
learning an adaptive state space abstraction and a communication protocol among
agents participating in navigation tasks. The goal is to come up with an
adaptive abstractor that significantly reduces the size of the state space to
be explored, without degradation in the policy performance. Simulation results
show that the proposed method reaches a better policy, in terms of achievable
rewards, resulting in fewer training iterations compared to the case where raw
states or fixed state abstraction are used. Moreover, it is shown that a
communication protocol emerges during training which enables the agents to
learn better policies within fewer training iterations.
Related papers
- DCMAC: Demand-aware Customized Multi-Agent Communication via Upper Bound Training [9.068971933560416]
We propose a Demand-aware Customized Multi-Agent Communication protocol, which use an upper bound training to obtain the ideal policy.
Experimental results reveal that DCMAC significantly outperforms the baseline algorithms in both unconstrained and communication constrained scenarios.
arXiv Detail & Related papers (2024-09-11T09:23:27Z) - Learning Multi-Agent Communication from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
Our proposed approach, CommFormer, efficiently optimize the communication graph and concurrently refines architectural parameters through gradient descent in an end-to-end manner.
arXiv Detail & Related papers (2024-05-14T12:40:25Z) - From Explicit Communication to Tacit Cooperation:A Novel Paradigm for
Cooperative MARL [14.935456456463731]
We propose a novel paradigm that facilitates a gradual shift from explicit communication to tacit cooperation.
In the initial training stage, we promote cooperation by sharing relevant information among agents.
We then combine the explicitly communicated information with the reconstructed information to obtain mixed information.
arXiv Detail & Related papers (2023-04-28T06:56:07Z) - On the Role of Emergent Communication for Social Learning in Multi-Agent
Reinforcement Learning [0.0]
Social learning uses cues from experts to align heterogeneous policies, reduce sample complexity, and solve partially observable tasks.
This paper proposes an unsupervised method based on the information bottleneck to capture both referential complexity and task-specific utility.
arXiv Detail & Related papers (2023-02-28T03:23:27Z) - Path Design and Resource Management for NOMA enhanced Indoor Intelligent
Robots [58.980293789967575]
A communication enabled indoor intelligent robots (IRs) service framework is proposed.
Lego modeling method is proposed, which can deterministically describe the indoor layout and channel state.
The investigated radio map is invoked as a virtual environment to train the reinforcement learning agent.
arXiv Detail & Related papers (2020-11-23T21:45:01Z) - Language-guided Navigation via Cross-Modal Grounding and Alternate
Adversarial Learning [66.9937776799536]
The emerging vision-and-language navigation (VLN) problem aims at learning to navigate an agent to the target location in unseen photo-realistic environments.
The main challenges of VLN arise mainly from two aspects: first, the agent needs to attend to the meaningful paragraphs of the language instruction corresponding to the dynamically-varying visual environments.
We propose a cross-modal grounding module to equip the agent with a better ability to track the correspondence between the textual and visual modalities.
arXiv Detail & Related papers (2020-11-22T09:13:46Z) - Networked Multi-Agent Reinforcement Learning with Emergent Communication [18.47483427884452]
Multi-Agent Reinforcement Learning (MARL) methods find optimal policies for agents that operate in the presence of other learning agents.
One way to coordinate is by learning to communicate with each other.
Can the agents develop a language while learning to perform a common task?
arXiv Detail & Related papers (2020-04-06T16:13:23Z) - Learning to cooperate: Emergent communication in multi-agent navigation [49.11609702016523]
We show that agents performing a cooperative navigation task learn an interpretable communication protocol.
An analysis of the agents' policies reveals that emergent signals spatially cluster the state space.
Using populations of agents, we show that the emergent protocol has basic compositional structure.
arXiv Detail & Related papers (2020-04-02T16:03:17Z) - Learning Structured Communication for Multi-agent Reinforcement Learning [104.64584573546524]
This work explores the large-scale multi-agent communication mechanism under a multi-agent reinforcement learning (MARL) setting.
We propose a novel framework termed as Learning Structured Communication (LSC) by using a more flexible and efficient communication topology.
arXiv Detail & Related papers (2020-02-11T07:19:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.