Neural Distillation as a State Representation Bottleneck in
Reinforcement Learning
- URL: http://arxiv.org/abs/2210.02224v1
- Date: Wed, 5 Oct 2022 13:00:39 GMT
- Title: Neural Distillation as a State Representation Bottleneck in
Reinforcement Learning
- Authors: Valentin Guillet, Dennis G. Wilson, Carlos Aguilar-Melchor, Emmanuel
Rachelson
- Abstract summary: We argue that distillation can be used to learn a state representation displaying favorable characteristics.
We first evaluate these criteria and verify the contribution of distillation on state representation on a toy environment based on the standard inverted pendulum problem.
- Score: 4.129225533930966
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Learning a good state representation is a critical skill when dealing with
multiple tasks in Reinforcement Learning as it allows for transfer and better
generalization between tasks. However, defining what constitute a useful
representation is far from simple and there is so far no standard method to
find such an encoding. In this paper, we argue that distillation -- a process
that aims at imitating a set of given policies with a single neural network --
can be used to learn a state representation displaying favorable
characteristics. In this regard, we define three criteria that measure
desirable features of a state encoding: the ability to select important
variables in the input space, the ability to efficiently separate states
according to their corresponding optimal action, and the robustness of the
state encoding on new tasks. We first evaluate these criteria and verify the
contribution of distillation on state representation on a toy environment based
on the standard inverted pendulum problem, before extending our analysis on
more complex visual tasks from the Atari and Procgen benchmarks.
Related papers
- Learning telic-controllable state representations [3.072340427031969]
We present a novel computational framework for state representation learning in bounded agents.
Our work advances a unified theoretical perspective on goal-directed state representation learning in natural and artificial agents.
arXiv Detail & Related papers (2024-06-20T16:38:25Z) - Locally Constrained Representations in Reinforcement Learning [5.865719902445064]
The success of Reinforcement Learning heavily relies on the ability to learn robust representations from the observations of the environment.
In most cases, the representations learned purely by the reinforcement learning loss can differ vastly across states depending on how the value functions change.
We propose locally constrained representations, where an auxiliary loss forces the state representations to be predictable by the representations of the neighboring states.
arXiv Detail & Related papers (2022-09-20T03:36:39Z) - Visual processing in context of reinforcement learning [0.0]
This thesis introduces three different representation learning algorithms that have access to different subsets of the data sources that traditional RL algorithms use.
We conclude that including unsupervised representation learning in RL problem-solving pipelines can speed up learning.
arXiv Detail & Related papers (2022-08-26T09:30:51Z) - Action Quality Assessment with Temporal Parsing Transformer [84.1272079121699]
Action Quality Assessment (AQA) is important for action understanding and resolving the task poses unique challenges due to subtle visual differences.
We propose a temporal parsing transformer to decompose the holistic feature into temporal part-level representations.
Our proposed method outperforms prior work on three public AQA benchmarks by a considerable margin.
arXiv Detail & Related papers (2022-07-19T13:29:05Z) - Value-Consistent Representation Learning for Data-Efficient
Reinforcement Learning [105.70602423944148]
We propose a novel method, called value-consistent representation learning (VCR), to learn representations that are directly related to decision-making.
Instead of aligning this imagined state with a real state returned by the environment, VCR applies a $Q$-value head on both states and obtains two distributions of action values.
It has been demonstrated that our methods achieve new state-of-the-art performance for search-free RL algorithms.
arXiv Detail & Related papers (2022-06-25T03:02:25Z) - Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon
Reasoning [120.38381203153159]
Reinforcement learning can train policies that effectively perform complex tasks.
For long-horizon tasks, the performance of these methods degrades with horizon, often necessitating reasoning over and composing lower-level skills.
We propose Value Function Spaces: a simple approach that produces such a representation by using the value functions corresponding to each lower-level skill.
arXiv Detail & Related papers (2021-11-04T22:46:16Z) - Learning Markov State Abstractions for Deep Reinforcement Learning [17.34529517221924]
We introduce a novel set of conditions and prove that they are sufficient for learning a Markov abstract state representation.
We then describe a practical training procedure that combines inverse model estimation and temporal contrastive learning.
Our approach learns representations that capture the underlying structure of the domain and lead to improved sample efficiency.
arXiv Detail & Related papers (2021-06-08T14:12:36Z) - Conditional Meta-Learning of Linear Representations [57.90025697492041]
Standard meta-learning for representation learning aims to find a common representation to be shared across multiple tasks.
In this work we overcome this issue by inferring a conditioning function, mapping the tasks' side information into a representation tailored to the task at hand.
We propose a meta-algorithm capable of leveraging this advantage in practice.
arXiv Detail & Related papers (2021-03-30T12:02:14Z) - Towards Learning Controllable Representations of Physical Systems [9.088303226909279]
Learned representations of dynamical systems reduce dimensionality, potentially supporting downstream reinforcement learning (RL)
We consider the relationship between the true state and the corresponding representations, proposing that ideally each representation corresponds to a unique state.
These metrics are shown to predict reinforcement learning performance in a simulated peg-in-hole task when comparing variants of autoencoder-based representations.
arXiv Detail & Related papers (2020-11-16T17:15:57Z) - Learning Invariant Representations for Reinforcement Learning without
Reconstruction [98.33235415273562]
We study how representation learning can accelerate reinforcement learning from rich observations, such as images, without relying either on domain knowledge or pixel-reconstruction.
Bisimulation metrics quantify behavioral similarity between states in continuous MDPs.
We demonstrate the effectiveness of our method at disregarding task-irrelevant information using modified visual MuJoCo tasks.
arXiv Detail & Related papers (2020-06-18T17:59:35Z) - Fairness by Learning Orthogonal Disentangled Representations [50.82638766862974]
We propose a novel disentanglement approach to invariant representation problem.
We enforce the meaningful representation to be agnostic to sensitive information by entropy.
The proposed approach is evaluated on five publicly available datasets.
arXiv Detail & Related papers (2020-03-12T11:09:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.