KnowRU: Knowledge Reusing via Knowledge Distillation in Multi-agent
Reinforcement Learning
- URL: http://arxiv.org/abs/2103.14891v1
- Date: Sat, 27 Mar 2021 12:38:01 GMT
- Title: KnowRU: Knowledge Reusing via Knowledge Distillation in Multi-agent
Reinforcement Learning
- Authors: Zijian Gao, Kele Xu, Bo Ding, Huaimin Wang, Yiying Li, Hongda Jia
- Abstract summary: Deep Reinforcement Learning (RL) algorithms have achieved dramatically progress in the multi-agent area.
To alleviate this problem, efficient leveraging of the historical experience is essential.
We propose a method, named "KnowRU" for knowledge reusing.
- Score: 16.167201058368303
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, deep Reinforcement Learning (RL) algorithms have achieved
dramatically progress in the multi-agent area. However, training the
increasingly complex tasks would be time-consuming and resources-exhausting. To
alleviate this problem, efficient leveraging the historical experience is
essential, which is under-explored in previous studies as most of the exiting
methods may fail to achieve this goal in a continuously variational system due
to their complicated design and environmental dynamics. In this paper, we
propose a method, named "KnowRU" for knowledge reusing which can be easily
deployed in the majority of the multi-agent reinforcement learning algorithms
without complicated hand-coded design. We employ the knowledge distillation
paradigm to transfer the knowledge among agents with the goal to accelerate the
training phase for new tasks, while improving the asymptotic performance of
agents. To empirically demonstrate the robustness and effectiveness of KnowRU,
we perform extensive experiments on state-of-the-art multi-agent reinforcement
learning (MARL) algorithms on collaborative and competitive scenarios. The
results show that KnowRU can outperform the recently reported methods, which
emphasizes the importance of the proposed knowledge reusing for MARL.
Related papers
- From Novice to Expert: LLM Agent Policy Optimization via Step-wise Reinforcement Learning [62.54484062185869]
We introduce StepAgent, which utilizes step-wise reward to optimize the agent's reinforcement learning process.
We propose implicit-reward and inverse reinforcement learning techniques to facilitate agent reflection and policy adjustment.
arXiv Detail & Related papers (2024-11-06T10:35:11Z) - Synergistic Multi-Agent Framework with Trajectory Learning for Knowledge-Intensive Tasks [44.42989163847349]
Large Language Models (LLMs) have led to significant breakthroughs in various natural language processing tasks.
generating factually consistent responses in knowledge-intensive scenarios remains a challenge.
This paper introduces SMART, a novel multi-agent framework that leverages external knowledge to enhance the interpretability and factual consistency of LLM-generated responses.
arXiv Detail & Related papers (2024-07-13T13:58:24Z) - WESE: Weak Exploration to Strong Exploitation for LLM Agents [95.6720931773781]
This paper proposes a novel approach, Weak Exploration to Strong Exploitation (WESE) to enhance LLM agents in solving open-world interactive tasks.
WESE involves decoupling the exploration and exploitation process, employing a cost-effective weak agent to perform exploration tasks for global knowledge.
A knowledge graph-based strategy is then introduced to store the acquired knowledge and extract task-relevant knowledge, enhancing the stronger agent in success rate and efficiency for the exploitation task.
arXiv Detail & Related papers (2024-04-11T03:31:54Z) - Enabling Multi-Agent Transfer Reinforcement Learning via Scenario
Independent Representation [0.7366405857677227]
Multi-Agent Reinforcement Learning (MARL) algorithms are widely adopted in tackling complex tasks that require collaboration and competition among agents.
We introduce a novel framework that enables transfer learning for MARL through unifying various state spaces into fixed-size inputs.
We show significant enhancements in multi-agent learning performance using maneuvering skills learned from other scenarios compared to agents learning from scratch.
arXiv Detail & Related papers (2024-02-13T02:48:18Z) - ExpeL: LLM Agents Are Experiential Learners [60.54312035818746]
We introduce the Experiential Learning (ExpeL) agent to allow learning from agent experiences without requiring parametric updates.
Our agent autonomously gathers experiences and extracts knowledge using natural language from a collection of training tasks.
At inference, the agent recalls its extracted insights and past experiences to make informed decisions.
arXiv Detail & Related papers (2023-08-20T03:03:34Z) - Rethinking Learning Dynamics in RL using Adversarial Networks [79.56118674435844]
We present a learning mechanism for reinforcement learning of closely related skills parameterized via a skill embedding space.
The main contribution of our work is to formulate an adversarial training regime for reinforcement learning with the help of entropy-regularized policy gradient formulation.
arXiv Detail & Related papers (2022-01-27T19:51:09Z) - Relational Experience Replay: Continual Learning by Adaptively Tuning
Task-wise Relationship [54.73817402934303]
We propose Experience Continual Replay (ERR), a bi-level learning framework to adaptively tune task-wise to achieve a better stability plasticity' tradeoff.
ERR can consistently improve the performance of all baselines and surpass current state-of-the-art methods.
arXiv Detail & Related papers (2021-12-31T12:05:22Z) - Learning from Guided Play: A Scheduled Hierarchical Approach for
Improving Exploration in Adversarial Imitation Learning [7.51557557629519]
We present Learning from Guided Play (LfGP), a framework in which we leverage expert demonstrations of, in addition to a main task, multiple auxiliary tasks.
This affords many benefits: learning efficiency is improved for main tasks with challenging bottleneck transitions, expert data becomes reusable between tasks, and transfer learning through the reuse of learned auxiliary task models becomes possible.
arXiv Detail & Related papers (2021-12-16T14:58:08Z) - KnowSR: Knowledge Sharing among Homogeneous Agents in Multi-agent
Reinforcement Learning [16.167201058368303]
We present an adaptation method of the majority of multi-agent reinforcement learning (MARL) algorithms called KnowSR.
We employ the idea of knowledge distillation (KD) to share knowledge among agents to shorten the training phase.
To empirically demonstrate the robustness and effectiveness of KnowSR, we performed extensive experiments on state-of-the-art MARL algorithms in collaborative and competitive scenarios.
arXiv Detail & Related papers (2021-05-25T02:19:41Z) - Meta-Reinforcement Learning Robust to Distributional Shift via Model
Identification and Experience Relabeling [126.69933134648541]
We present a meta-reinforcement learning algorithm that is both efficient and extrapolates well when faced with out-of-distribution tasks at test time.
Our method is based on a simple insight: we recognize that dynamics models can be adapted efficiently and consistently with off-policy data.
arXiv Detail & Related papers (2020-06-12T13:34:46Z) - A new Potential-Based Reward Shaping for Reinforcement Learning Agent [0.0]
The proposed method extracts knowledge from episodes' cumulative rewards.
The results indicate an improvement in the learning process in both the single-task and the multi-task reinforcement learner agents.
arXiv Detail & Related papers (2019-02-17T10:34:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.