Self-supervised Reinforcement Learning with Independently Controllable
Subgoals
- URL: http://arxiv.org/abs/2109.04150v1
- Date: Thu, 9 Sep 2021 10:21:02 GMT
- Title: Self-supervised Reinforcement Learning with Independently Controllable
Subgoals
- Authors: Andrii Zadaianchuk, Georg Martius, Fanny Yang
- Abstract summary: Self-supervised agents set their own goals by exploiting the structure in the environment.
Some of them were applied to learn basic manipulation skills in compositional multi-object environments.
We propose a novel self-supervised agent that estimates relations between environment components and uses them to independently control different parts of the environment state.
- Score: 20.29444813790076
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To successfully tackle challenging manipulation tasks, autonomous agents must
learn a diverse set of skills and how to combine them. Recently,
self-supervised agents that set their own abstract goals by exploiting the
discovered structure in the environment were shown to perform well on many
different tasks. In particular, some of them were applied to learn basic
manipulation skills in compositional multi-object environments. However, these
methods learn skills without taking the dependencies between objects into
account. Thus, the learned skills are difficult to combine in realistic
environments. We propose a novel self-supervised agent that estimates relations
between environment components and uses them to independently control different
parts of the environment state. In addition, the estimated relations between
objects can be used to decompose a complex goal into a compatible sequence of
subgoals. We show that, by using this framework, an agent can efficiently and
automatically learn manipulation tasks in multi-object environments with
different relations between objects.
Related papers
- Exploiting Contextual Structure to Generate Useful Auxiliary Tasks [12.236546713167945]
Reinforcement learning requires interaction with an environment, which is expensive for robots.
We propose an approach that maximizes experience reuse while learning to solve a given task by generating and simultaneously learning useful auxiliary tasks.
arXiv Detail & Related papers (2023-03-09T05:11:30Z) - Dexterous Manipulation from Images: Autonomous Real-World RL via Substep
Guidance [71.36749876465618]
We describe a system for vision-based dexterous manipulation that provides a "programming-free" approach for users to define new tasks.
Our system includes a framework for users to define a final task and intermediate sub-tasks with image examples.
experimental results with a four-finger robotic hand learning multi-stage object manipulation tasks directly in the real world.
arXiv Detail & Related papers (2022-12-19T22:50:40Z) - Inferring Versatile Behavior from Demonstrations by Matching Geometric
Descriptors [72.62423312645953]
Humans intuitively solve tasks in versatile ways, varying their behavior in terms of trajectory-based planning and for individual steps.
Current Imitation Learning algorithms often only consider unimodal expert demonstrations and act in a state-action-based setting.
Instead, we combine a mixture of movement primitives with a distribution matching objective to learn versatile behaviors that match the expert's behavior and versatility.
arXiv Detail & Related papers (2022-10-17T16:42:59Z) - Autonomous Open-Ended Learning of Tasks with Non-Stationary
Interdependencies [64.0476282000118]
Intrinsic motivations have proven to generate a task-agnostic signal to properly allocate the training time amongst goals.
While the majority of works in the field of intrinsically motivated open-ended learning focus on scenarios where goals are independent from each other, only few of them studied the autonomous acquisition of interdependent tasks.
In particular, we first deepen the analysis of a previous system, showing the importance of incorporating information about the relationships between tasks at a higher level of the architecture.
Then we introduce H-GRAIL, a new system that extends the previous one by adding a new learning layer to store the autonomously acquired sequences
arXiv Detail & Related papers (2022-05-16T10:43:01Z) - Policy Architectures for Compositional Generalization in Control [71.61675703776628]
We introduce a framework for modeling entity-based compositional structure in tasks.
Our policies are flexible and can be trained end-to-end without requiring any action primitives.
arXiv Detail & Related papers (2022-03-10T06:44:24Z) - Multi-Agent Imitation Learning with Copulas [102.27052968901894]
Multi-agent imitation learning aims to train multiple agents to perform tasks from demonstrations by learning a mapping between observations and actions.
In this paper, we propose to use copula, a powerful statistical tool for capturing dependence among random variables, to explicitly model the correlation and coordination in multi-agent systems.
Our proposed model is able to separately learn marginals that capture the local behavioral patterns of each individual agent, as well as a copula function that solely and fully captures the dependence structure among agents.
arXiv Detail & Related papers (2021-07-10T03:49:41Z) - Self-supervised Visual Reinforcement Learning with Object-centric
Representations [11.786249372283562]
We propose to use object-centric representations as a modular and structured observation space.
We show that the structure in the representations in combination with goal-conditioned attention policies helps the autonomous agent to discover and learn useful skills.
arXiv Detail & Related papers (2020-11-29T14:55:09Z) - Autonomous learning of multiple, context-dependent tasks [1.1470070927586016]
In complex environments, the same task might need a set of different skills to be solved.
We propose a novel open-ended learning robot architecture, C-GRAIL, that solves the two challenges in an integrated fashion.
The architecture is tested in a simulated robotic environment involving a robot that autonomously learns to reach relevant target objects.
arXiv Detail & Related papers (2020-11-27T17:25:36Z) - CausalWorld: A Robotic Manipulation Benchmark for Causal Structure and
Transfer Learning [138.40338621974954]
CausalWorld is a benchmark for causal structure and transfer learning in a robotic manipulation environment.
Tasks consist of constructing 3D shapes from a given set of blocks - inspired by how children learn to build complex structures.
arXiv Detail & Related papers (2020-10-08T23:01:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.