Skill-based Multi-objective Reinforcement Learning of Industrial Robot
Tasks with Planning and Knowledge Integration
- URL: http://arxiv.org/abs/2203.10033v1
- Date: Fri, 18 Mar 2022 16:03:27 GMT
- Title: Skill-based Multi-objective Reinforcement Learning of Industrial Robot
Tasks with Planning and Knowledge Integration
- Authors: Matthias Mayr, Faseeh Ahmad, Konstantinos Chatzilygeroudis, Luigi
Nardi, Volker Krueger
- Abstract summary: We introduce an approach that provides a combination of task-level planning with targeted learning of scenario-specific parameters for skill-based systems.
We demonstrate the efficacy and versatility of our approach by learning skill parameters for two different contact-rich tasks.
- Score: 0.4949816699298335
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In modern industrial settings with small batch sizes it should be easy to set
up a robot system for a new task. Strategies exist, e.g. the use of skills, but
when it comes to handling forces and torques, these systems often fall short.
We introduce an approach that provides a combination of task-level planning
with targeted learning of scenario-specific parameters for skill-based systems.
We propose the following pipeline: (1) the user provides a task goal in the
planning language PDDL, (2) a plan (i.e., a sequence of skills) is generated
and the learnable parameters of the skills are automatically identified. An
operator then chooses (3) reward functions and hyperparameters for the learning
process. Two aspects of our methodology are critical: (a) learning is tightly
integrated with a knowledge framework to support symbolic planning and to
provide priors for learning, (b) using multi-objective optimization. This can
help to balance key performance indicators (KPIs) such as safety and task
performance since they can often affect each other. We adopt a multi-objective
Bayesian optimization approach and learn entirely in simulation. We demonstrate
the efficacy and versatility of our approach by learning skill parameters for
two different contact-rich tasks. We show their successful execution on a real
7-DOF KUKA-iiwa manipulator and outperform the manual parameterization by human
robot operators.
Related papers
- SPIRE: Synergistic Planning, Imitation, and Reinforcement Learning for Long-Horizon Manipulation [58.14969377419633]
We propose spire, a system that decomposes tasks into smaller learning subproblems and second combines imitation and reinforcement learning to maximize their strengths.
We find that spire outperforms prior approaches that integrate imitation learning, reinforcement learning, and planning by 35% to 50% in average task performance.
arXiv Detail & Related papers (2024-10-23T17:42:07Z) - Learning Reward for Robot Skills Using Large Language Models via Self-Alignment [11.639973274337274]
Large Language Models (LLM) contain valuable task-related knowledge that can potentially aid in the learning of reward functions.
We propose a method to learn rewards more efficiently in the absence of humans.
arXiv Detail & Related papers (2024-05-12T04:57:43Z) - Practice Makes Perfect: Planning to Learn Skill Parameter Policies [34.51008914846429]
In this work, we focus on the active learning problem of choosing which skills to practice to maximize expected future task success.
We propose that the robot should estimate the competence of each skill, extrapolate the competence, and situate the skill in the task distribution through competence-aware planning.
This approach is implemented within a fully autonomous system where the robot repeatedly plans, practices, and learns without any environment resets.
arXiv Detail & Related papers (2024-02-22T23:58:26Z) - Mastering Robot Manipulation with Multimodal Prompts through Pretraining and Multi-task Fine-tuning [49.92517970237088]
We tackle the problem of training a robot to understand multimodal prompts.
This type of task poses a major challenge to robots' capability to understand the interconnection and complementarity between vision and language signals.
We introduce an effective framework that learns a policy to perform robot manipulation with multimodal prompts.
arXiv Detail & Related papers (2023-10-14T22:24:58Z) - Active Task Randomization: Learning Robust Skills via Unsupervised
Generation of Diverse and Feasible Tasks [37.73239471412444]
We introduce Active Task Randomization (ATR), an approach that learns robust skills through the unsupervised generation of training tasks.
ATR selects suitable tasks, which consist of an initial environment state and manipulation goal, for learning robust skills by balancing the diversity and feasibility of the tasks.
We demonstrate that the learned skills can be composed by a task planner to solve unseen sequential manipulation problems based on visual inputs.
arXiv Detail & Related papers (2022-11-11T11:24:55Z) - Learning Neuro-Symbolic Skills for Bilevel Planning [63.388694268198655]
Decision-making is challenging in robotics environments with continuous object-centric states, continuous actions, long horizons, and sparse feedback.
Hierarchical approaches, such as task and motion planning (TAMP), address these challenges by decomposing decision-making into two or more levels of abstraction.
Our main contribution is a method for learning parameterized polices in combination with operators and samplers.
arXiv Detail & Related papers (2022-06-21T19:01:19Z) - Meta-Learning Parameterized Skills [12.845774297648738]
We propose a novel skill-learning algorithm that aims to learn transferable parameterized skills and synthesize them into a new action space.
We empirically demonstrate that the proposed algorithms enable an agent to solve a set of difficult long-horizon (obstacle-course and robot manipulation) tasks.
arXiv Detail & Related papers (2022-06-07T21:30:58Z) - Autonomous Open-Ended Learning of Tasks with Non-Stationary
Interdependencies [64.0476282000118]
Intrinsic motivations have proven to generate a task-agnostic signal to properly allocate the training time amongst goals.
While the majority of works in the field of intrinsically motivated open-ended learning focus on scenarios where goals are independent from each other, only few of them studied the autonomous acquisition of interdependent tasks.
In particular, we first deepen the analysis of a previous system, showing the importance of incorporating information about the relationships between tasks at a higher level of the architecture.
Then we introduce H-GRAIL, a new system that extends the previous one by adding a new learning layer to store the autonomously acquired sequences
arXiv Detail & Related papers (2022-05-16T10:43:01Z) - Combining Modular Skills in Multitask Learning [149.8001096811708]
A modular design encourages neural models to disentangle and recombine different facets of knowledge to generalise more systematically to new tasks.
In this work, we assume each task is associated with a subset of latent discrete skills from a (potentially small) inventory.
We find that the modular design of a network significantly increases sample efficiency in reinforcement learning and few-shot generalisation in supervised learning.
arXiv Detail & Related papers (2022-02-28T16:07:19Z) - MT-Opt: Continuous Multi-Task Robotic Reinforcement Learning at Scale [103.7609761511652]
We show how a large-scale collective robotic learning system can acquire a repertoire of behaviors simultaneously.
New tasks can be continuously instantiated from previously learned tasks.
We train and evaluate our system on a set of 12 real-world tasks with data collected from 7 robots.
arXiv Detail & Related papers (2021-04-16T16:38:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.