Relate to Predict: Towards Task-Independent Knowledge Representations
for Reinforcement Learning
- URL: http://arxiv.org/abs/2212.05298v1
- Date: Sat, 10 Dec 2022 13:33:56 GMT
- Title: Relate to Predict: Towards Task-Independent Knowledge Representations
for Reinforcement Learning
- Authors: Thomas Schn\"urer, Malte Probst, Horst-Michael Gross
- Abstract summary: Reinforcement Learning can enable agents to learn complex tasks.
It is difficult to interpret the knowledge and reuse it across tasks.
In this paper, we introduce an inductive bias for explicit object-centered knowledge separation.
We show that the degree of explicitness in knowledge separation correlates with faster learning, better accuracy, better generalization, and better interpretability.
- Score: 11.245432408899092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement Learning (RL) can enable agents to learn complex tasks.
However, it is difficult to interpret the knowledge and reuse it across tasks.
Inductive biases can address such issues by explicitly providing generic yet
useful decomposition that is otherwise difficult or expensive to learn
implicitly. For example, object-centered approaches decompose a high
dimensional observation into individual objects. Expanding on this, we utilize
an inductive bias for explicit object-centered knowledge separation that
provides further decomposition into semantic representations and dynamics
knowledge. For this, we introduce a semantic module that predicts an objects'
semantic state based on its context. The resulting affordance-like object state
can then be used to enrich perceptual object representations. With a minimal
setup and an environment that enables puzzle-like tasks, we demonstrate the
feasibility and benefits of this approach. Specifically, we compare three
different methods of integrating semantic representations into a model-based RL
architecture. Our experiments show that the degree of explicitness in knowledge
separation correlates with faster learning, better accuracy, better
generalization, and better interpretability.
Related papers
- A Probabilistic Model Behind Self-Supervised Learning [53.64989127914936]
In self-supervised learning (SSL), representations are learned via an auxiliary task without annotated labels.
We present a generative latent variable model for self-supervised learning.
We show that several families of discriminative SSL, including contrastive methods, induce a comparable distribution over representations.
arXiv Detail & Related papers (2024-02-02T13:31:17Z) - Recognizing Unseen Objects via Multimodal Intensive Knowledge Graph
Propagation [68.13453771001522]
We propose a multimodal intensive ZSL framework that matches regions of images with corresponding semantic embeddings.
We conduct extensive experiments and evaluate our model on large-scale real-world data.
arXiv Detail & Related papers (2023-06-14T13:07:48Z) - On the Generalization of Learned Structured Representations [5.1398743023989555]
We study methods that learn, with little or no supervision, representations of unstructured data that capture its hidden structure.
The second part of this thesis focuses on object-centric representations, which capture the compositional structure of the input in terms of symbol-like entities.
arXiv Detail & Related papers (2023-04-25T17:14:36Z) - Robust and Controllable Object-Centric Learning through Energy-based
Models [95.68748828339059]
ours is a conceptually simple and general approach to learning object-centric representations through an energy-based model.
We show that ours can be easily integrated into existing architectures and can effectively extract high-quality object-centric representations.
arXiv Detail & Related papers (2022-10-11T15:11:15Z) - On Neural Architecture Inductive Biases for Relational Tasks [76.18938462270503]
We introduce a simple architecture based on similarity-distribution scores which we name Compositional Network generalization (CoRelNet)
We find that simple architectural choices can outperform existing models in out-of-distribution generalizations.
arXiv Detail & Related papers (2022-06-09T16:24:01Z) - Self-Supervised Visual Representation Learning with Semantic Grouping [50.14703605659837]
We tackle the problem of learning visual representations from unlabeled scene-centric data.
We propose contrastive learning from data-driven semantic slots, namely SlotCon, for joint semantic grouping and representation learning.
arXiv Detail & Related papers (2022-05-30T17:50:59Z) - An Empirical Investigation of Representation Learning for Imitation [76.48784376425911]
Recent work in vision, reinforcement learning, and NLP has shown that auxiliary representation learning objectives can reduce the need for large amounts of expensive, task-specific data.
We propose a modular framework for constructing representation learning algorithms, then use our framework to evaluate the utility of representation learning for imitation.
arXiv Detail & Related papers (2022-05-16T11:23:42Z) - Object Pursuit: Building a Space of Objects via Discriminative Weight
Generation [23.85039747700698]
We propose a framework to continuously learn object-centric representations for visual learning and understanding.
We leverage interactions to sample diverse variations of an object and the corresponding training signals while learning the object-centric representations.
We perform an extensive study of the key features of the proposed framework and analyze the characteristics of the learned representations.
arXiv Detail & Related papers (2021-12-15T08:25:30Z) - Generalization and Robustness Implications in Object-Centric Learning [23.021791024676986]
In this paper, we train state-of-the-art unsupervised models on five common multi-object datasets.
From our experimental study, we find object-centric representations to be generally useful for downstream tasks.
arXiv Detail & Related papers (2021-07-01T17:51:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.