Representation Learning in Deep RL via Discrete Information Bottleneck
- URL: http://arxiv.org/abs/2212.13835v2
- Date: Wed, 31 May 2023 03:14:34 GMT
- Title: Representation Learning in Deep RL via Discrete Information Bottleneck
- Authors: Riashat Islam, Hongyu Zang, Manan Tomar, Aniket Didolkar, Md Mofijul
Islam, Samin Yeasar Arnob, Tariq Iqbal, Xin Li, Anirudh Goyal, Nicolas Heess,
Alex Lamb
- Abstract summary: We study how information bottlenecks can be used to construct latent states efficiently in the presence of task-irrelevant information.
We propose architectures that utilize variational and discrete information bottlenecks, coined as RepDIB, to learn structured factorized representations.
- Score: 39.375822469572434
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Several self-supervised representation learning methods have been proposed
for reinforcement learning (RL) with rich observations. For real-world
applications of RL, recovering underlying latent states is crucial,
particularly when sensory inputs contain irrelevant and exogenous information.
In this work, we study how information bottlenecks can be used to construct
latent states efficiently in the presence of task-irrelevant information. We
propose architectures that utilize variational and discrete information
bottlenecks, coined as RepDIB, to learn structured factorized representations.
Exploiting the expressiveness bought by factorized representations, we
introduce a simple, yet effective, bottleneck that can be integrated with any
existing self-supervised objective for RL. We demonstrate this across several
online and offline RL benchmarks, along with a real robot arm task, where we
find that compressed representations with RepDIB can lead to strong performance
improvements, as the learned bottlenecks help predict only the relevant state
while ignoring irrelevant information.
Related papers
- DEAR: Disentangled Environment and Agent Representations for Reinforcement Learning without Reconstruction [4.813546138483559]
Reinforcement Learning (RL) algorithms can learn robotic control tasks from visual observations, but they often require a large amount of data.
In this paper, we explore how the agent's knowledge of its shape can improve the sample efficiency of visual RL methods.
We propose a novel method, Disentangled Environment and Agent Representations, that uses the segmentation mask of the agent as supervision.
arXiv Detail & Related papers (2024-06-30T09:15:21Z) - Accelerating exploration and representation learning with offline
pre-training [52.6912479800592]
We show that exploration and representation learning can be improved by separately learning two different models from a single offline dataset.
We show that learning a state representation using noise-contrastive estimation and a model of auxiliary reward can significantly improve the sample efficiency on the challenging NetHack benchmark.
arXiv Detail & Related papers (2023-03-31T18:03:30Z) - Agent-Controller Representations: Principled Offline RL with Rich
Exogenous Information [49.06422815335159]
Learning to control an agent from data collected offline is vital for real-world applications of reinforcement learning (RL)
This paper introduces offline RL benchmarks offering the ability to study this problem.
We find that contemporary representation learning techniques can fail on datasets where the noise is a complex and time dependent process.
arXiv Detail & Related papers (2022-10-31T22:12:48Z) - Efficient entity-based reinforcement learning [3.867363075280544]
We propose to combine recent advances in set representations with slot attention and graph neural networks to process structured data.
We show that it can improve training time and robustness significantly, and demonstrate their potential to handle structured as well as purely visual domains.
arXiv Detail & Related papers (2022-06-06T19:02:39Z) - INFOrmation Prioritization through EmPOWERment in Visual Model-Based RL [90.06845886194235]
We propose a modified objective for model-based reinforcement learning (RL)
We integrate a term inspired by variational empowerment into a state-space model based on mutual information.
We evaluate the approach on a suite of vision-based robot control tasks with natural video backgrounds.
arXiv Detail & Related papers (2022-04-18T23:09:23Z) - Learning Temporally-Consistent Representations for Data-Efficient
Reinforcement Learning [3.308743964406687]
$k$-Step Latent (KSL) is a representation learning method that enforces temporal consistency of representations.
KSL produces encoders that generalize better to new tasks unseen during training.
arXiv Detail & Related papers (2021-10-11T00:16:43Z) - Reinforcement Learning with Prototypical Representations [114.35801511501639]
Proto-RL is a self-supervised framework that ties representation learning with exploration through prototypical representations.
These prototypes simultaneously serve as a summarization of the exploratory experience of an agent as well as a basis for representing observations.
This enables state-of-the-art downstream policy learning on a set of difficult continuous control tasks.
arXiv Detail & Related papers (2021-02-22T18:56:34Z) - Provably Efficient Causal Reinforcement Learning with Confounded
Observational Data [135.64775986546505]
We study how to incorporate the dataset (observational data) collected offline, which is often abundantly available in practice, to improve the sample efficiency in the online setting.
We propose the deconfounded optimistic value iteration (DOVI) algorithm, which incorporates the confounded observational data in a provably efficient manner.
arXiv Detail & Related papers (2020-06-22T14:49:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.