Entity Divider with Language Grounding in Multi-Agent Reinforcement
Learning
- URL: http://arxiv.org/abs/2210.13942v1
- Date: Tue, 25 Oct 2022 11:53:52 GMT
- Title: Entity Divider with Language Grounding in Multi-Agent Reinforcement
Learning
- Authors: Ziluo Ding, Wanpeng Zhang, Junpeng Yue, Xiangjun Wang, Tiejun Huang,
and Zongqing Lu
- Abstract summary: We investigate the use of natural language to drive the generalization of policies in multi-agent settings.
We propose a novel framework for language grounding in multi-agent reinforcement learning, entity divider (EnDi)
EnDi enables agents to independently learn subgoal division at the entity level and act in the environment based on the associated entities.
- Score: 28.619845209653274
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We investigate the use of natural language to drive the generalization of
policies in multi-agent settings. Unlike single-agent settings, the
generalization of policies should also consider the influence of other agents.
Besides, with the increasing number of entities in multi-agent settings, more
agent-entity interactions are needed for language grounding, and the enormous
search space could impede the learning process. Moreover, given a simple
general instruction,e.g., beating all enemies, agents are required to decompose
it into multiple subgoals and figure out the right one to focus on. Inspired by
previous work, we try to address these issues at the entity level and propose a
novel framework for language grounding in multi-agent reinforcement learning,
entity divider (EnDi). EnDi enables agents to independently learn subgoal
division at the entity level and act in the environment based on the associated
entities. The subgoal division is regularized by opponent modeling to avoid
subgoal conflicts and promote coordinated strategies. Empirically, EnDi
demonstrates the strong generalization ability to unseen games with new
dynamics and expresses the superiority over existing methods.
Related papers
- Policy Learning with a Language Bottleneck [65.99843627646018]
Policy Learning with a Language Bottleneck (PLLBB) is a framework enabling AI agents to generate linguistic rules.
PLLBB alternates between a rule generation step guided by language models, and an update step where agents learn new policies guided by rules.
In a two-player communication game, a maze solving task, and two image reconstruction tasks, we show thatPLLBB agents are not only able to learn more interpretable and generalizable behaviors, but can also share the learned rules with human users.
arXiv Detail & Related papers (2024-05-07T08:40:21Z) - AgentVerse: Facilitating Multi-Agent Collaboration and Exploring
Emergent Behaviors [93.38830440346783]
We propose a multi-agent framework framework that can collaboratively adjust its composition as a greater-than-the-sum-of-its-parts system.
Our experiments demonstrate that framework framework can effectively deploy multi-agent groups that outperform a single agent.
In view of these behaviors, we discuss some possible strategies to leverage positive ones and mitigate negative ones for improving the collaborative potential of multi-agent groups.
arXiv Detail & Related papers (2023-08-21T16:47:11Z) - Multi-Level Compositional Reasoning for Interactive Instruction
Following [24.581542880280203]
Multi-level Compositional Reasoning Agent (MCR-Agent)
At the highest level, we infer a sequence of human-interpretable subgoals to be executed based on language instructions by a high-level policy composition controller.
At the middle level, we discriminatively control the agent's navigation by a master policy by alternating between a navigation policy and various independent interaction policies.
At the lowest level, we infer manipulation actions with the corresponding object masks using the appropriate interaction policy.
arXiv Detail & Related papers (2023-08-18T08:38:28Z) - Retroformer: Retrospective Large Language Agents with Policy Gradient Optimization [103.70896967077294]
This paper introduces a principled framework for reinforcing large language agents by learning a retrospective model.
Our proposed agent architecture learns from rewards across multiple environments and tasks, for fine-tuning a pre-trained language model.
Experimental results on various tasks demonstrate that the language agents improve over time.
arXiv Detail & Related papers (2023-08-04T06:14:23Z) - Multi-agent Deep Covering Skill Discovery [50.812414209206054]
We propose Multi-agent Deep Covering Option Discovery, which constructs the multi-agent options through minimizing the expected cover time of the multiple agents' joint state space.
Also, we propose a novel framework to adopt the multi-agent options in the MARL process.
We show that the proposed algorithm can effectively capture the agent interactions with the attention mechanism, successfully identify multi-agent options, and significantly outperforms prior works using single-agent options or no options.
arXiv Detail & Related papers (2022-10-07T00:40:59Z) - ALMA: Hierarchical Learning for Composite Multi-Agent Tasks [21.556661319375255]
We introduce ALMA, a general learning method for taking advantage of structured tasks.
ALMA simultaneously learns a high-level subtask allocation policy and low-level agent policies.
We demonstrate that ALMA learns sophisticated coordination behavior in a number of challenging environments.
arXiv Detail & Related papers (2022-05-27T19:12:23Z) - Interpretable Reinforcement Learning with Multilevel Subgoal Discovery [77.34726150561087]
We propose a novel Reinforcement Learning model for discrete environments.
In the model, an agent learns information about environment in the form of probabilistic rules.
No reward function is required for learning; an agent only needs to be given a primary goal to achieve.
arXiv Detail & Related papers (2022-02-15T14:04:44Z) - Learning to Ground Multi-Agent Communication with Autoencoders [43.22048280036316]
Communication requires a common language, a lingua franca, between agents.
We demonstrate a simple way to ground language in learned representations.
We find that a standard representation learning algorithm is sufficient for arriving at a grounded common language.
arXiv Detail & Related papers (2021-10-28T17:57:26Z) - A Policy Gradient Algorithm for Learning to Learn in Multiagent
Reinforcement Learning [47.154539984501895]
We propose a novel meta-multiagent policy gradient theorem that accounts for the non-stationary policy dynamics inherent to multiagent learning settings.
This is achieved by modeling our gradient updates to consider both an agent's own non-stationary policy dynamics and the non-stationary policy dynamics of other agents in the environment.
arXiv Detail & Related papers (2020-10-31T22:50:21Z) - The Emergence of Adversarial Communication in Multi-Agent Reinforcement
Learning [6.18778092044887]
Many real-world problems require the coordination of multiple autonomous agents.
Recent work has shown the promise of Graph Neural Networks (GNNs) to learn explicit communication strategies that enable complex multi-agent coordination.
We show how a single self-interested agent is capable of learning highly manipulative communication strategies that allows it to significantly outperform a cooperative team of agents.
arXiv Detail & Related papers (2020-08-06T12:48:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.