Transforming task representations to perform novel tasks
- URL: http://arxiv.org/abs/2005.04318v3
- Date: Tue, 6 Oct 2020 18:35:56 GMT
- Title: Transforming task representations to perform novel tasks
- Authors: Andrew K. Lampinen and James L. McClelland
- Abstract summary: An important aspect of intelligence is the ability to adapt to a novel task without any direct experience (zero-shot)
We propose a general computational framework for adapting to novel tasks based on their relationship to prior tasks.
- Score: 12.008469282323492
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An important aspect of intelligence is the ability to adapt to a novel task
without any direct experience (zero-shot), based on its relationship to
previous tasks. Humans can exhibit this cognitive flexibility. By contrast,
models that achieve superhuman performance in specific tasks often fail to
adapt to even slight task alterations. To address this, we propose a general
computational framework for adapting to novel tasks based on their relationship
to prior tasks. We begin by learning vector representations of tasks. To adapt
to new tasks, we propose meta-mappings, higher-order tasks that transform basic
task representations. We demonstrate the effectiveness of this framework across
a wide variety of tasks and computational paradigms, ranging from regression to
image classification and reinforcement learning. We compare to both human
adaptability and language-based approaches to zero-shot learning. Across these
domains, meta-mapping is successful, often achieving 80-90% performance,
without any data, on a novel task, even when the new task directly contradicts
prior experience. We further show that meta-mapping can not only generalize to
new tasks via learned relationships, but can also generalize using novel
relationships unseen during training. Finally, using meta-mapping as a starting
point can dramatically accelerate later learning on a new task, and reduce
learning time and cumulative error substantially. Our results provide insight
into a possible computational basis of intelligent adaptability and offer a
possible framework for modeling cognitive flexibility and building more
flexible artificial intelligence systems.
Related papers
- Generalizing to New Tasks via One-Shot Compositional Subgoals [23.15624959305799]
The ability to generalize to previously unseen tasks with little to no supervision is a key challenge in modern machine learning research.
We introduce CASE which attempts to address these issues by training an Imitation Learning agent using adaptive "near future" subgoals.
Our experiments show that the proposed approach consistently outperforms the previous state-of-the-art compositional Imitation Learning approach by 30%.
arXiv Detail & Related papers (2022-05-16T14:30:11Z) - Multi-Agent Policy Transfer via Task Relationship Modeling [28.421365805638953]
We try to discover and exploit common structures among tasks for more efficient transfer.
We propose to learn effect-based task representations as a common space of tasks, using an alternatively fixed training scheme.
As a result, the proposed method can help transfer learned cooperation knowledge to new tasks after training on a few source tasks.
arXiv Detail & Related papers (2022-03-09T01:49:21Z) - Towards More Generalizable One-shot Visual Imitation Learning [81.09074706236858]
A general-purpose robot should be able to master a wide range of tasks and quickly learn a novel one by leveraging past experiences.
One-shot imitation learning (OSIL) approaches this goal by training an agent with (pairs of) expert demonstrations.
We push for a higher level of generalization ability by investigating a more ambitious multi-task setup.
arXiv Detail & Related papers (2021-10-26T05:49:46Z) - Adaptive Transfer Learning on Graph Neural Networks [4.233435459239147]
Graph neural networks (GNNs) are widely used to learn a powerful representation of graph-structured data.
Recent work demonstrates that transferring knowledge from self-supervised tasks to downstream tasks could further improve graph representation.
We propose a new transfer learning paradigm on GNNs which could effectively leverage self-supervised tasks as auxiliary tasks to help the target task.
arXiv Detail & Related papers (2021-07-19T11:46:28Z) - Latent Skill Planning for Exploration and Transfer [49.25525932162891]
In this paper, we investigate how these two approaches can be integrated into a single reinforcement learning agent.
We leverage the idea of partial amortization for fast adaptation at test time.
We demonstrate the benefits of our design decisions across a suite of challenging locomotion tasks.
arXiv Detail & Related papers (2020-11-27T18:40:03Z) - Parrot: Data-Driven Behavioral Priors for Reinforcement Learning [79.32403825036792]
We propose a method for pre-training behavioral priors that can capture complex input-output relationships observed in successful trials.
We show how this learned prior can be used for rapidly learning new tasks without impeding the RL agent's ability to try out novel behaviors.
arXiv Detail & Related papers (2020-11-19T18:47:40Z) - Importance Weighted Policy Learning and Adaptation [89.46467771037054]
We study a complementary approach which is conceptually simple, general, modular and built on top of recent improvements in off-policy learning.
The framework is inspired by ideas from the probabilistic inference literature and combines robust off-policy learning with a behavior prior.
Our approach achieves competitive adaptation performance on hold-out tasks compared to meta reinforcement learning baselines and can scale to complex sparse-reward scenarios.
arXiv Detail & Related papers (2020-09-10T14:16:58Z) - Meta-Reinforcement Learning Robust to Distributional Shift via Model
Identification and Experience Relabeling [126.69933134648541]
We present a meta-reinforcement learning algorithm that is both efficient and extrapolates well when faced with out-of-distribution tasks at test time.
Our method is based on a simple insight: we recognize that dynamics models can be adapted efficiently and consistently with off-policy data.
arXiv Detail & Related papers (2020-06-12T13:34:46Z) - Representation Ensembling for Synergistic Lifelong Learning with
Quasilinear Complexity [17.858926093389737]
In lifelong learning, data are used to improve performance not only on the current task, but also on previously encountered, and as yet unencountered tasks.
Our key insight is that we can synergistically ensemble representations -- that were learned independently on disparate tasks -- to enable both forward and backward transfer.
arXiv Detail & Related papers (2020-04-27T16:16:30Z) - A Boolean Task Algebra for Reinforcement Learning [14.731788603429774]
We formalise the logical composition of tasks as a Boolean algebra.
We show that by learning goal-oriented value functions, an agent can solve new tasks with no further learning.
arXiv Detail & Related papers (2020-01-06T04:46:25Z) - Automated Relational Meta-learning [95.02216511235191]
We propose an automated relational meta-learning framework that automatically extracts the cross-task relations and constructs the meta-knowledge graph.
We conduct extensive experiments on 2D toy regression and few-shot image classification and the results demonstrate the superiority of ARML over state-of-the-art baselines.
arXiv Detail & Related papers (2020-01-03T07:02:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.