Improving Context-Based Meta-Reinforcement Learning with Self-Supervised
Trajectory Contrastive Learning
- URL: http://arxiv.org/abs/2103.06386v1
- Date: Wed, 10 Mar 2021 23:31:19 GMT
- Title: Improving Context-Based Meta-Reinforcement Learning with Self-Supervised
Trajectory Contrastive Learning
- Authors: Bernie Wang, Simon Xu, Kurt Keutzer, Yang Gao, Bichen Wu
- Abstract summary: We propose Trajectory Contrastive Learning to improve meta-training.
TCL trains a context encoder to predict whether two transition windows are sampled from the same trajectory.
It accelerates the training of context encoders and improves meta-training overall.
- Score: 32.112504515457445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Meta-reinforcement learning typically requires orders of magnitude more
samples than single task reinforcement learning methods. This is because
meta-training needs to deal with more diverse distributions and train extra
components such as context encoders. To address this, we propose a novel
self-supervised learning task, which we named Trajectory Contrastive Learning
(TCL), to improve meta-training. TCL adopts contrastive learning and trains a
context encoder to predict whether two transition windows are sampled from the
same trajectory. TCL leverages the natural hierarchical structure of
context-based meta-RL and makes minimal assumptions, allowing it to be
generally applicable to context-based meta-RL algorithms. It accelerates the
training of context encoders and improves meta-training overall. Experiments
show that TCL performs better or comparably than a strong meta-RL baseline in
most of the environments on both meta-RL MuJoCo (5 of 6) and Meta-World
benchmarks (44 out of 50).
Related papers
- AMAGO-2: Breaking the Multi-Task Barrier in Meta-Reinforcement Learning with Transformers [28.927809804613215]
We build upon recent advancements in Transformer-based (in-context) meta-RL.
We evaluate a simple yet scalable solution where both an agent's actor and critic objectives are converted to classification terms.
This design unlocks significant progress in online multi-task adaptation and memory problems without explicit task labels.
arXiv Detail & Related papers (2024-11-17T22:25:40Z) - ConML: A Universal Meta-Learning Framework with Task-Level Contrastive Learning [49.447777286862994]
ConML is a universal meta-learning framework that can be applied to various meta-learning algorithms.
We demonstrate that ConML integrates seamlessly with optimization-based, metric-based, and amortization-based meta-learning algorithms.
arXiv Detail & Related papers (2024-10-08T12:22:10Z) - Meta Reinforcement Learning with Successor Feature Based Context [51.35452583759734]
We propose a novel meta-RL approach that achieves competitive performance comparing to existing meta-RL algorithms.
Our method does not only learn high-quality policies for multiple tasks simultaneously but also can quickly adapt to new tasks with a small amount of training.
arXiv Detail & Related papers (2022-07-29T14:52:47Z) - Learning Action Translator for Meta Reinforcement Learning on
Sparse-Reward Tasks [56.63855534940827]
This work introduces a novel objective function to learn an action translator among training tasks.
We theoretically verify that the value of the transferred policy with the action translator can be close to the value of the source policy.
We propose to combine the action translator with context-based meta-RL algorithms for better data collection and more efficient exploration during meta-training.
arXiv Detail & Related papers (2022-07-19T04:58:06Z) - On the Effectiveness of Fine-tuning Versus Meta-reinforcement Learning [71.55412580325743]
We show that multi-task pretraining with fine-tuning on new tasks performs equally as well, or better, than meta-pretraining with meta test-time adaptation.
This is encouraging for future research, as multi-task pretraining tends to be simpler and computationally cheaper than meta-RL.
arXiv Detail & Related papers (2022-06-07T13:24:00Z) - MetaICL: Learning to Learn In Context [87.23056864536613]
We introduce MetaICL, a new meta-training framework for few-shot learning where a pretrained language model is tuned to do in-context learn-ing on a large set of training tasks.
We show that MetaICL approaches (and sometimes beats) the performance of models fully finetuned on the target task training data, and outperforms much bigger models with nearly 8x parameters.
arXiv Detail & Related papers (2021-10-29T17:42:08Z) - Towards Effective Context for Meta-Reinforcement Learning: an Approach
based on Contrastive Learning [33.19862944149082]
We propose a novel Meta-RL framework called CCM (Contrastive learning augmented Context-based Meta-RL)
We first focus on the contrastive nature behind different tasks and leverage it to train a compact and sufficient context encoder.
We derive a new information-gain-based objective which aims to collect informative trajectories in a few steps.
arXiv Detail & Related papers (2020-09-29T09:29:18Z) - Incremental Meta-Learning via Indirect Discriminant Alignment [118.61152684795178]
We develop a notion of incremental learning during the meta-training phase of meta-learning.
Our approach performs favorably at test time as compared to training a model with the full meta-training set.
arXiv Detail & Related papers (2020-02-11T01:39:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.