Active World Model Learning with Progress Curiosity
- URL: http://arxiv.org/abs/2007.07853v1
- Date: Wed, 15 Jul 2020 17:19:17 GMT
- Title: Active World Model Learning with Progress Curiosity
- Authors: Kuno Kim, Megumi Sano, Julian De Freitas, Nick Haber, Daniel Yamins
- Abstract summary: World models are self-supervised predictive models of how the world evolves.
In this work, we study how to design such a curiosity-driven Active World Model Learning system.
We propose an AWML system driven by $gamma$-Progress: a scalable and effective learning progress-based curiosity signal.
- Score: 12.077052764803163
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: World models are self-supervised predictive models of how the world evolves.
Humans learn world models by curiously exploring their environment, in the
process acquiring compact abstractions of high bandwidth sensory inputs, the
ability to plan across long temporal horizons, and an understanding of the
behavioral patterns of other agents. In this work, we study how to design such
a curiosity-driven Active World Model Learning (AWML) system. To do so, we
construct a curious agent building world models while visually exploring a 3D
physical environment rich with distillations of representative real-world
agents. We propose an AWML system driven by $\gamma$-Progress: a scalable and
effective learning progress-based curiosity signal. We show that
$\gamma$-Progress naturally gives rise to an exploration policy that directs
attention to complex but learnable dynamics in a balanced manner, thus
overcoming the "white noise problem". As a result, our $\gamma$-Progress-driven
controller achieves significantly higher AWML performance than baseline
controllers equipped with state-of-the-art exploration strategies such as
Random Network Distillation and Model Disagreement.
Related papers
- SOLD: Reinforcement Learning with Slot Object-Centric Latent Dynamics [16.020835290802548]
Slot-Attention for Object-centric Latent Dynamics is a novel algorithm that learns object-centric dynamics models from pixel inputs.
We demonstrate that the structured latent space not only improves model interpretability but also provides a valuable input space for behavior models to reason over.
Our results show that SOLD outperforms DreamerV3, a state-of-the-art model-based RL algorithm, across a range of benchmark robotic environments.
arXiv Detail & Related papers (2024-10-11T14:03:31Z) - Is Sora a World Simulator? A Comprehensive Survey on General World Models and Beyond [101.15395503285804]
General world models represent a crucial pathway toward achieving Artificial General Intelligence (AGI)
In this survey, we embark on a comprehensive exploration of the latest advancements in world models.
We examine challenges and limitations of world models, and discuss their potential future directions.
arXiv Detail & Related papers (2024-05-06T14:37:07Z) - STORM: Efficient Stochastic Transformer based World Models for
Reinforcement Learning [82.03481509373037]
Recently, model-based reinforcement learning algorithms have demonstrated remarkable efficacy in visual input environments.
We introduce Transformer-based wORld Model (STORM), an efficient world model architecture that combines strong modeling and generation capabilities.
Storm achieves a mean human performance of $126.7%$ on the Atari $100$k benchmark, setting a new record among state-of-the-art methods.
arXiv Detail & Related papers (2023-10-14T16:42:02Z) - HarmonyDream: Task Harmonization Inside World Models [93.07314830304193]
Model-based reinforcement learning (MBRL) holds the promise of sample-efficient learning.
We propose a simple yet effective approach, HarmonyDream, which automatically adjusts loss coefficients to maintain task harmonization.
arXiv Detail & Related papers (2023-09-30T11:38:13Z) - MoDem-V2: Visuo-Motor World Models for Real-World Robot Manipulation [29.349549097807916]
MoDem-V2 is capable of learning contact-rich manipulation directly in the uninstrumented real world.
We identify key ingredients for leveraging demonstrations in model learning while respecting real-world safety considerations.
arXiv Detail & Related papers (2023-09-25T15:51:29Z) - Predictive Experience Replay for Continual Visual Control and
Forecasting [62.06183102362871]
We present a new continual learning approach for visual dynamics modeling and explore its efficacy in visual control and forecasting.
We first propose the mixture world model that learns task-specific dynamics priors with a mixture of Gaussians, and then introduce a new training strategy to overcome catastrophic forgetting.
Our model remarkably outperforms the naive combinations of existing continual learning and visual RL algorithms on DeepMind Control and Meta-World benchmarks with continual visual control tasks.
arXiv Detail & Related papers (2023-03-12T05:08:03Z) - Bridging the Gap to Real-World Object-Centric Learning [66.55867830853803]
We show that reconstructing features from models trained in a self-supervised manner is a sufficient training signal for object-centric representations to arise in a fully unsupervised way.
Our approach, DINOSAUR, significantly out-performs existing object-centric learning models on simulated data.
arXiv Detail & Related papers (2022-09-29T15:24:47Z) - Masked World Models for Visual Control [90.13638482124567]
We introduce a visual model-based RL framework that decouples visual representation learning and dynamics learning.
We demonstrate that our approach achieves state-of-the-art performance on a variety of visual robotic tasks.
arXiv Detail & Related papers (2022-06-28T18:42:27Z) - Curious Exploration via Structured World Models Yields Zero-Shot Object
Manipulation [19.840186443344]
We propose to use structured world models to incorporate inductive biases in the control loop to achieve sample-efficient exploration.
Our method generates free-play behavior that starts to interact with objects early on and develops more complex behavior over time.
arXiv Detail & Related papers (2022-06-22T22:08:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.