Dreaming: Model-based Reinforcement Learning by Latent Imagination
without Reconstruction
- URL: http://arxiv.org/abs/2007.14535v2
- Date: Fri, 12 Mar 2021 03:56:41 GMT
- Title: Dreaming: Model-based Reinforcement Learning by Latent Imagination
without Reconstruction
- Authors: Masashi Okada, Tadahiro Taniguchi
- Abstract summary: We propose a decoder-free extension of Dreamer, a leading model-based reinforcement learning (MBRL) method from pixels.
We derive a likelihood-free and InfoMax objective of contrastive learning from the evidence lower bound of Dreamer.
In comparison to Dreamer and other recent model-free reinforcement learning methods, our newly devised Dreamer with InfoMax and without generative decoder (Dreaming) achieves the best scores on 5 difficult simulated robotics tasks.
- Score: 14.950054143767824
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the present paper, we propose a decoder-free extension of Dreamer, a
leading model-based reinforcement learning (MBRL) method from pixels. Dreamer
is a sample- and cost-efficient solution to robot learning, as it is used to
train latent state-space models based on a variational autoencoder and to
conduct policy optimization by latent trajectory imagination. However, this
autoencoding based approach often causes object vanishing, in which the
autoencoder fails to perceives key objects for solving control tasks, and thus
significantly limiting Dreamer's potential. This work aims to relieve this
Dreamer's bottleneck and enhance its performance by means of removing the
decoder. For this purpose, we firstly derive a likelihood-free and InfoMax
objective of contrastive learning from the evidence lower bound of Dreamer.
Secondly, we incorporate two components, (i) independent linear dynamics and
(ii) the random crop data augmentation, to the learning scheme so as to improve
the training performance. In comparison to Dreamer and other recent model-free
reinforcement learning methods, our newly devised Dreamer with InfoMax and
without generative decoder (Dreaming) achieves the best scores on 5 difficult
simulated robotics tasks, in which Dreamer suffers from object vanishing.
Related papers
- Dreaming is All You Need [6.706761900460658]
This research introduces two novel deep learning models, SleepNet and DreamNet, to strike a balance between exploration and precision.
SleepNet seamlessly integrates supervised learning with unsupervised sleep" stages using pre-trained encoder models.
DreamNet employs full encoder-decoder frameworks to reconstruct the hidden states, mimicking the human "dreaming" process.
arXiv Detail & Related papers (2024-09-03T06:04:39Z) - CURLing the Dream: Contrastive Representations for World Modeling in Reinforcement Learning [0.22615818641180724]
Curled-Dreamer is a novel reinforcement learning algorithm that integrates contrastive learning into the DreamerV3 framework.
Our experiments demonstrate that Curled-Dreamer consistently outperforms state-of-the-art algorithms.
arXiv Detail & Related papers (2024-08-11T14:13:22Z) - MuDreamer: Learning Predictive World Models without Reconstruction [58.0159270859475]
We present MuDreamer, a robust reinforcement learning agent that builds upon the DreamerV3 algorithm by learning a predictive world model without the need for reconstructing input signals.
Our method achieves comparable performance on the Atari100k benchmark while benefiting from faster training.
arXiv Detail & Related papers (2024-05-23T22:09:01Z) - Self-Improving Robots: End-to-End Autonomous Visuomotor Reinforcement
Learning [54.636562516974884]
In imitation and reinforcement learning, the cost of human supervision limits the amount of data that robots can be trained on.
In this work, we propose MEDAL++, a novel design for self-improving robotic systems.
The robot autonomously practices the task by learning to both do and undo the task, simultaneously inferring the reward function from the demonstrations.
arXiv Detail & Related papers (2023-03-02T18:51:38Z) - Masked World Models for Visual Control [90.13638482124567]
We introduce a visual model-based RL framework that decouples visual representation learning and dynamics learning.
We demonstrate that our approach achieves state-of-the-art performance on a variety of visual robotic tasks.
arXiv Detail & Related papers (2022-06-28T18:42:27Z) - DayDreamer: World Models for Physical Robot Learning [142.11031132529524]
Deep reinforcement learning is a common approach to robot learning but requires a large amount of trial and error to learn.
Many advances in robot learning rely on simulators.
In this paper, we apply Dreamer to 4 robots to learn online and directly in the real world, without simulators.
arXiv Detail & Related papers (2022-06-28T17:44:48Z) - What Language Model Architecture and Pretraining Objective Work Best for
Zero-Shot Generalization? [50.84738303888189]
We present a large-scale evaluation of modeling choices and their impact on zero-shot generalization.
We train models with over 5 billion parameters for more than 170 billion tokens.
We find that pretrained causal decoder models can be efficiently adapted into non-causal decoder models.
arXiv Detail & Related papers (2022-04-12T14:19:49Z) - DreamingV2: Reinforcement Learning with Discrete World Models without
Reconstruction [14.950054143767824]
The present paper proposes a novel reinforcement learning method with world models, DreamingV2.
DreamingV2 is a collaborative extension of DreamerV2 and Dreaming.
We believe DreamingV2 will be a reliable solution for robot learning since its discrete representation is suitable to describe discontinuous environments.
arXiv Detail & Related papers (2022-03-01T14:44:15Z) - Mastering Atari with Discrete World Models [61.7688353335468]
We introduce DreamerV2, a reinforcement learning agent that learns behaviors purely from predictions in the compact latent space of a powerful world model.
DreamerV2 constitutes the first agent that achieves human-level performance on the Atari benchmark of 55 tasks by learning behaviors inside a separately trained world model.
arXiv Detail & Related papers (2020-10-05T17:52:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.