Learning State Representations via Retracing in Reinforcement Learning
- URL: http://arxiv.org/abs/2111.12600v1
- Date: Wed, 24 Nov 2021 16:19:59 GMT
- Title: Learning State Representations via Retracing in Reinforcement Learning
- Authors: Changmin Yu, Dong Li, Jianye Hao, Jun Wang, Neil Burgess
- Abstract summary: Learning via retracing is a self-supervised approach for learning the state representation for reinforcement learning tasks.
We introduce Cycle-Consistency World Model (CCWM), a concrete instantiation of learning via retracing.
We show that CCWM achieves state-of-the-art performance in terms of sample efficiency and performance.
- Score: 25.755855290244103
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose learning via retracing, a novel self-supervised approach for
learning the state representation (and the associated dynamics model) for
reinforcement learning tasks. In addition to the predictive (reconstruction)
supervision in the forward direction, we propose to include `"retraced"
transitions for representation/model learning, by enforcing the
cycle-consistency constraint between the original and retraced states, hence
improve upon the sample efficiency of learning. Moreover, learning via
retracing explicitly propagates information about future transitions backward
for inferring previous states, thus facilitates stronger representation
learning. We introduce Cycle-Consistency World Model (CCWM), a concrete
instantiation of learning via retracing implemented under existing model-based
reinforcement learning framework. Additionally we propose a novel adaptive
"truncation" mechanism for counteracting the negative impacts brought by the
"irreversible" transitions such that learning via retracing can be maximally
effective. Through extensive empirical studies on continuous control
benchmarks, we demonstrates that CCWM achieves state-of-the-art performance in
terms of sample efficiency and asymptotic performance.
Related papers
- Temporal-Difference Variational Continual Learning [89.32940051152782]
A crucial capability of Machine Learning models in real-world applications is the ability to continuously learn new tasks.
In Continual Learning settings, models often struggle to balance learning new tasks with retaining previous knowledge.
We propose new learning objectives that integrate the regularization effects of multiple previous posterior estimations.
arXiv Detail & Related papers (2024-10-10T10:58:41Z) - iQRL -- Implicitly Quantized Representations for Sample-efficient Reinforcement Learning [24.684363928059113]
We propose an efficient representation learning method using only a self-supervised latent-state consistency loss.
We achieve high performance and prevent representation collapse by quantizing the latent representation.
Our method, named iQRL: implicitly Quantized Reinforcement Learning, is straightforward, compatible with any model-free RL algorithm.
arXiv Detail & Related papers (2024-06-04T18:15:44Z) - Relaxed Contrastive Learning for Federated Learning [48.96253206661268]
We propose a novel contrastive learning framework to address the challenges of data heterogeneity in federated learning.
Our framework outperforms all existing federated learning approaches by huge margins on the standard benchmarks.
arXiv Detail & Related papers (2024-01-10T04:55:24Z) - Variance-Covariance Regularization Improves Representation Learning [28.341622247252705]
We adapt a self-supervised learning regularization technique to supervised learning contexts, introducing Variance-Covariance Regularization (VCReg)
We demonstrate that VCReg significantly enhances transfer learning for images and videos, achieving state-of-the-art performance across numerous tasks and datasets.
In summary, VCReg offers a universally applicable regularization framework that significantly advances transfer learning and highlights the connection between gradient starvation, neural collapse, and feature transferability.
arXiv Detail & Related papers (2023-06-23T05:01:02Z) - ICICLE: Interpretable Class Incremental Continual Learning [35.105786309067895]
Interpretable Class-InCremental LEarning (ICICLE) is an exemplar-free approach that adopts a prototypical part-based approach.
Our experimental results demonstrate that ICICLE reduces the interpretability concept drift and outperforms the existing exemplar-free methods of common class-incremental learning.
arXiv Detail & Related papers (2023-03-14T11:31:45Z) - Latent Variable Representation for Reinforcement Learning [131.03944557979725]
It remains unclear theoretically and empirically how latent variable models may facilitate learning, planning, and exploration to improve the sample efficiency of model-based reinforcement learning.
We provide a representation view of the latent variable models for state-action value functions, which allows both tractable variational learning algorithm and effective implementation of the optimism/pessimism principle.
In particular, we propose a computationally efficient planning algorithm with UCB exploration by incorporating kernel embeddings of latent variable models.
arXiv Detail & Related papers (2022-12-17T00:26:31Z) - ReIL: A Framework for Reinforced Intervention-based Imitation Learning [3.0846824529023387]
We introduce Reinforced Intervention-based Learning (ReIL), a framework consisting of a general intervention-based learning algorithm and a multi-task imitation learning model.
Experimental results from real world mobile robot navigation challenges indicate that ReIL learns rapidly from sparse supervisor corrections without suffering deterioration in performance.
arXiv Detail & Related papers (2022-03-29T09:30:26Z) - Co$^2$L: Contrastive Continual Learning [69.46643497220586]
Recent breakthroughs in self-supervised learning show that such algorithms learn visual representations that can be transferred better to unseen tasks.
We propose a rehearsal-based continual learning algorithm that focuses on continually learning and maintaining transferable representations.
arXiv Detail & Related papers (2021-06-28T06:14:38Z) - Return-Based Contrastive Representation Learning for Reinforcement
Learning [126.7440353288838]
We propose a novel auxiliary task that forces the learnt representations to discriminate state-action pairs with different returns.
Our algorithm outperforms strong baselines on complex tasks in Atari games and DeepMind Control suite.
arXiv Detail & Related papers (2021-02-22T13:04:18Z) - Guided Variational Autoencoder for Disentanglement Learning [79.02010588207416]
We propose an algorithm, guided variational autoencoder (Guided-VAE), that is able to learn a controllable generative model by performing latent representation disentanglement learning.
We design an unsupervised strategy and a supervised strategy in Guided-VAE and observe enhanced modeling and controlling capability over the vanilla VAE.
arXiv Detail & Related papers (2020-04-02T20:49:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.