Masked Autoencoding for Scalable and Generalizable Decision Making
- URL: http://arxiv.org/abs/2211.12740v2
- Date: Sat, 27 May 2023 09:16:38 GMT
- Title: Masked Autoencoding for Scalable and Generalizable Decision Making
- Authors: Fangchen Liu, Hao Liu, Aditya Grover, Pieter Abbeel
- Abstract summary: MaskDP is a simple and scalable self-supervised pretraining method for reinforcement learning and behavioral cloning.
We find that a MaskDP model gains the capability of zero-shot transfer to new BC tasks, such as single and multiple goal reaching.
- Score: 93.84855114717062
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We are interested in learning scalable agents for reinforcement learning that
can learn from large-scale, diverse sequential data similar to current large
vision and language models. To this end, this paper presents masked decision
prediction (MaskDP), a simple and scalable self-supervised pretraining method
for reinforcement learning (RL) and behavioral cloning (BC). In our MaskDP
approach, we employ a masked autoencoder (MAE) to state-action trajectories,
wherein we randomly mask state and action tokens and reconstruct the missing
data. By doing so, the model is required to infer masked-out states and actions
and extract information about dynamics. We find that masking different
proportions of the input sequence significantly helps with learning a better
model that generalizes well to multiple downstream tasks. In our empirical
study, we find that a MaskDP model gains the capability of zero-shot transfer
to new BC tasks, such as single and multiple goal reaching, and it can
zero-shot infer skills from a few example transitions. In addition, MaskDP
transfers well to offline RL and shows promising scaling behavior w.r.t. to
model size. It is amenable to data-efficient finetuning, achieving competitive
results with prior methods based on autoregressive pretraining.
Related papers
- Self-Supervised Neuron Segmentation with Multi-Agent Reinforcement
Learning [53.00683059396803]
Mask image model (MIM) has been widely used due to its simplicity and effectiveness in recovering original information from masked images.
We propose a decision-based MIM that utilizes reinforcement learning (RL) to automatically search for optimal image masking ratio and masking strategy.
Our approach has a significant advantage over alternative self-supervised methods on the task of neuron segmentation.
arXiv Detail & Related papers (2023-10-06T10:40:46Z) - CL-MAE: Curriculum-Learned Masked Autoencoders [49.24994655813455]
We propose a curriculum learning approach that updates the masking strategy to continually increase the complexity of the self-supervised reconstruction task.
We train our Curriculum-Learned Masked Autoencoder (CL-MAE) on ImageNet and show that it exhibits superior representation learning capabilities compared to MAE.
arXiv Detail & Related papers (2023-08-31T09:13:30Z) - Improving Masked Autoencoders by Learning Where to Mask [65.89510231743692]
Masked image modeling is a promising self-supervised learning method for visual data.
We present AutoMAE, a framework that uses Gumbel-Softmax to interlink an adversarially-trained mask generator and a mask-guided image modeling process.
In our experiments, AutoMAE is shown to provide effective pretraining models on standard self-supervised benchmarks and downstream tasks.
arXiv Detail & Related papers (2023-03-12T05:28:55Z) - RePreM: Representation Pre-training with Masked Model for Reinforcement
Learning [28.63696288537304]
We propose a masked model for pre-training in RL, RePreM, which trains the encoder combined with transformer blocks to predict the masked states or actions in a trajectory.
We show that RePreM scales well with dataset size, dataset quality, and the scale of the encoder, which indicates its potential towards big RL models.
arXiv Detail & Related papers (2023-03-03T02:04:14Z) - Exploring Target Representations for Masked Autoencoders [78.57196600585462]
We show that a careful choice of the target representation is unnecessary for learning good representations.
We propose a multi-stage masked distillation pipeline and use a randomly model as the teacher.
A proposed method to perform masked knowledge distillation with bootstrapped teachers (dBOT) outperforms previous self-supervised methods by nontrivial margins.
arXiv Detail & Related papers (2022-09-08T16:55:19Z) - Extreme Masking for Learning Instance and Distributed Visual
Representations [50.152264456036114]
The paper presents a scalable approach for learning distributed representations over individual tokens and a holistic instance representation simultaneously.
We use self-attention blocks to represent distributed tokens, followed by cross-attention blocks to aggregate the holistic instance.
Our model, named ExtreMA, follows the plain BYOL approach where the instance representation from the unmasked subset is trained to predict that from the intact input.
arXiv Detail & Related papers (2022-06-09T17:59:43Z) - Training Neural Networks with Fixed Sparse Masks [19.58969772430058]
Recent work has shown that it is possible to update only a small subset of the model's parameters during training.
We show that it is possible to induce a fixed sparse mask on the model's parameters that selects a subset to update over many iterations.
arXiv Detail & Related papers (2021-11-18T18:06:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.