Value Explicit Pretraining for Learning Transferable Representations
- URL: http://arxiv.org/abs/2312.12339v2
- Date: Thu, 7 Mar 2024 10:07:02 GMT
- Title: Value Explicit Pretraining for Learning Transferable Representations
- Authors: Kiran Lekkala, Henghui Bao, Sumedh Sontakke, Laurent Itti
- Abstract summary: We propose a method that learns generalizable representations for transfer reinforcement learning.
We learn new tasks that share similar objectives as previously learned tasks, by learning an encoder for objective-conditioned representations.
Experiments using a realistic navigation simulator and Atari benchmark show that the pretrained encoder produced by our method outperforms current SoTA pretraining methods.
- Score: 11.069853883599102
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose Value Explicit Pretraining (VEP), a method that learns
generalizable representations for transfer reinforcement learning. VEP enables
learning of new tasks that share similar objectives as previously learned
tasks, by learning an encoder for objective-conditioned representations,
irrespective of appearance changes and environment dynamics. To pre-train the
encoder from a sequence of observations, we use a self-supervised contrastive
loss that results in learning temporally smooth representations. VEP learns to
relate states across different tasks based on the Bellman return estimate that
is reflective of task progress. Experiments using a realistic navigation
simulator and Atari benchmark show that the pretrained encoder produced by our
method outperforms current SoTA pretraining methods on the ability to
generalize to unseen tasks. VEP achieves up to a 2 times improvement in rewards
on Atari and visual navigation, and up to a 3 times improvement in sample
efficiency. For videos of policy performance visit our
https://sites.google.com/view/value-explicit-pretraining/
Related papers
- Vision Language Models are In-Context Value Learners [89.29486557646624]
We present Generative Value Learning (GVL), a universal value function estimator that leverages the world knowledge embedded in vision-language models (VLMs) to predict task progress.
Without any robot or task specific training, GVL can in-context zero-shot and few-shot predict effective values for more than 300 distinct real-world tasks.
arXiv Detail & Related papers (2024-11-07T09:17:50Z) - Pre-trained Visual Dynamics Representations for Efficient Policy Learning [33.62440075940917]
We propose Pre-trained Visual Dynamics Representations (PVDR) to bridge the domain gap between videos and downstream tasks for efficient policy learning.
The pre-trained visual dynamics representations capture the visual dynamics prior knowledge in the videos.
This abstract prior knowledge can be readily adapted to downstream tasks and aligned with executable actions through online adaptation.
arXiv Detail & Related papers (2024-11-05T15:18:02Z) - Investigating Pre-Training Objectives for Generalization in Vision-Based Reinforcement Learning [41.1883663165909]
We introduce the Atari Pre-training Benchmark (Atari-PB), which pre-trains a ResNet-50 model on 10 million transitions from 50 Atari games.
Our experiments show that pre-training objectives focused on learning task-agnostic features enhance generalization across different environments.
In contrast, objectives focused on learning task-specific knowledge improve performance in environments similar to the pre-training dataset but not in varied ones.
arXiv Detail & Related papers (2024-06-10T06:06:38Z) - What Makes Pre-Trained Visual Representations Successful for Robust
Manipulation? [57.92924256181857]
We find that visual representations designed for manipulation and control tasks do not necessarily generalize under subtle changes in lighting and scene texture.
We find that emergent segmentation ability is a strong predictor of out-of-distribution generalization among ViT models.
arXiv Detail & Related papers (2023-11-03T18:09:08Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - Pretraining the Vision Transformer using self-supervised methods for
vision based Deep Reinforcement Learning [0.0]
We study pretraining a Vision Transformer using several state-of-the-art self-supervised methods and assess the quality of the learned representations.
Our results show that all methods are effective in learning useful representations and avoiding representational collapse.
The encoder pretrained with the temporal order verification task shows the best results across all experiments.
arXiv Detail & Related papers (2022-09-22T10:18:59Z) - Reinforcement Learning with Action-Free Pre-Training from Videos [95.25074614579646]
We introduce a framework that learns representations useful for understanding the dynamics via generative pre-training on videos.
Our framework significantly improves both final performances and sample-efficiency of vision-based reinforcement learning.
arXiv Detail & Related papers (2022-03-25T19:44:09Z) - Learning Invariant Representations for Reinforcement Learning without
Reconstruction [98.33235415273562]
We study how representation learning can accelerate reinforcement learning from rich observations, such as images, without relying either on domain knowledge or pixel-reconstruction.
Bisimulation metrics quantify behavioral similarity between states in continuous MDPs.
We demonstrate the effectiveness of our method at disregarding task-irrelevant information using modified visual MuJoCo tasks.
arXiv Detail & Related papers (2020-06-18T17:59:35Z) - Pre-training Text Representations as Meta Learning [113.3361289756749]
We introduce a learning algorithm which directly optimize model's ability to learn text representations for effective learning of downstream tasks.
We show that there is an intrinsic connection between multi-task pre-training and model-agnostic meta-learning with a sequence of meta-train steps.
arXiv Detail & Related papers (2020-04-12T09:05:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.