Multi-Agent Policy Transfer via Task Relationship Modeling
- URL: http://arxiv.org/abs/2203.04482v1
- Date: Wed, 9 Mar 2022 01:49:21 GMT
- Title: Multi-Agent Policy Transfer via Task Relationship Modeling
- Authors: Rongjun Qin, Feng Chen, Tonghan Wang, Lei Yuan, Xiaoran Wu, Zongzhang
Zhang, Chongjie Zhang, Yang Yu
- Abstract summary: We try to discover and exploit common structures among tasks for more efficient transfer.
We propose to learn effect-based task representations as a common space of tasks, using an alternatively fixed training scheme.
As a result, the proposed method can help transfer learned cooperation knowledge to new tasks after training on a few source tasks.
- Score: 28.421365805638953
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Team adaptation to new cooperative tasks is a hallmark of human intelligence,
which has yet to be fully realized in learning agents. Previous work on
multi-agent transfer learning accommodate teams of different sizes, heavily
relying on the generalization ability of neural networks for adapting to unseen
tasks. We believe that the relationship among tasks provides the key
information for policy adaptation. In this paper, we try to discover and
exploit common structures among tasks for more efficient transfer, and propose
to learn effect-based task representations as a common space of tasks, using an
alternatively fixed training scheme. We demonstrate that the task
representation can capture the relationship among tasks, and can generalize to
unseen tasks. As a result, the proposed method can help transfer learned
cooperation knowledge to new tasks after training on a few source tasks. We
also find that fine-tuning the transferred policies help solve tasks that are
hard to learn from scratch.
Related papers
- TaskWeb: Selecting Better Source Tasks for Multi-task NLP [76.03221609799931]
Knowing task relationships via pairwise task transfer improves choosing one or more source tasks that help to learn a new target task.
We use TaskWeb to estimate the benefit of using a source task for learning a new target task, and to choose a subset of helpful training tasks for multi-task training.
Our method improves overall rankings and top-k precision of source tasks by 10% and 38%, respectively.
arXiv Detail & Related papers (2023-05-22T17:27:57Z) - Task Compass: Scaling Multi-task Pre-training with Task Prefix [122.49242976184617]
Existing studies show that multi-task learning with large-scale supervised tasks suffers from negative effects across tasks.
We propose a task prefix guided multi-task pre-training framework to explore the relationships among tasks.
Our model can not only serve as the strong foundation backbone for a wide range of tasks but also be feasible as a probing tool for analyzing task relationships.
arXiv Detail & Related papers (2022-10-12T15:02:04Z) - Transferring Knowledge for Reinforcement Learning in Contact-Rich
Manipulation [10.219833196479142]
We address the challenge of transferring knowledge within a family of similar tasks by leveraging multiple skill priors.
Our method learns a latent action space representing the skill embedding from demonstrated trajectories for each prior task.
We have evaluated our method on a set of peg-in-hole insertion tasks and demonstrate better generalization to new tasks that have never been encountered during training.
arXiv Detail & Related papers (2022-09-19T10:31:13Z) - Learning Task Embeddings for Teamwork Adaptation in Multi-Agent
Reinforcement Learning [13.468555224407764]
We show that a team of agents is able to adapt to novel tasks when provided with task embeddings.
We propose three MATE training paradigms: independent MATE, centralised MATE, and mixed MATE.
We show that the embeddings learned by MATE identify tasks and provide useful information which agents leverage during adaptation to novel tasks.
arXiv Detail & Related papers (2022-07-05T18:23:20Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z) - Modular Adaptive Policy Selection for Multi-Task Imitation Learning
through Task Division [60.232542918414985]
Multi-task learning often suffers from negative transfer, sharing information that should be task-specific.
This is done by using proto-policies as modules to divide the tasks into simple sub-behaviours that can be shared.
We also demonstrate its ability to autonomously divide the tasks into both shared and task-specific sub-behaviours.
arXiv Detail & Related papers (2022-03-28T15:53:17Z) - Efficiently Identifying Task Groupings for Multi-Task Learning [55.80489920205404]
Multi-task learning can leverage information learned by one task to benefit the training of other tasks.
We suggest an approach to select which tasks should train together in multi-task learning models.
Our method determines task groupings in a single training run by co-training all tasks together and quantifying the effect to which one task's gradient would affect another task's loss.
arXiv Detail & Related papers (2021-09-10T02:01:43Z) - Adaptive Transfer Learning on Graph Neural Networks [4.233435459239147]
Graph neural networks (GNNs) are widely used to learn a powerful representation of graph-structured data.
Recent work demonstrates that transferring knowledge from self-supervised tasks to downstream tasks could further improve graph representation.
We propose a new transfer learning paradigm on GNNs which could effectively leverage self-supervised tasks as auxiliary tasks to help the target task.
arXiv Detail & Related papers (2021-07-19T11:46:28Z) - Adaptive Policy Transfer in Reinforcement Learning [9.594432031144715]
We introduce a principled mechanism that can "Adapt-to-Learn", that is adapt the source policy to learn to solve a target task.
We show that the presented method learns to seamlessly combine learning from adaptation and exploration and leads to a robust policy transfer algorithm.
arXiv Detail & Related papers (2021-05-10T22:42:03Z) - Measuring and Harnessing Transference in Multi-Task Learning [58.48659733262734]
Multi-task learning can leverage information learned by one task to benefit the training of other tasks.
We analyze the dynamics of information transfer, or transference, across tasks throughout training.
arXiv Detail & Related papers (2020-10-29T08:25:43Z) - Transforming task representations to perform novel tasks [12.008469282323492]
An important aspect of intelligence is the ability to adapt to a novel task without any direct experience (zero-shot)
We propose a general computational framework for adapting to novel tasks based on their relationship to prior tasks.
arXiv Detail & Related papers (2020-05-08T23:41:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.