RMP2: A Structured Composable Policy Class for Robot Learning
- URL: http://arxiv.org/abs/2103.05922v1
- Date: Wed, 10 Mar 2021 08:28:38 GMT
- Title: RMP2: A Structured Composable Policy Class for Robot Learning
- Authors: Anqi Li, Ching-An Cheng, M. Asif Rana, Man Xie, Karl Van Wyk, Nathan
Ratliff, Byron Boots
- Abstract summary: We consider the problem of learning motion policies for acceleration-based robotics systems with a structured policy class specified by RMPflow.
RMPflow is a multi-task control framework that has been successfully applied in many robotics problems.
We re-examine the message passing algorithm of RMPflow called RMP2 and propose a more efficient algorithm to compute RMPflow policies.
- Score: 36.35483747142448
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of learning motion policies for acceleration-based
robotics systems with a structured policy class specified by RMPflow. RMPflow
is a multi-task control framework that has been successfully applied in many
robotics problems. Using RMPflow as a structured policy class in learning has
several benefits, such as sufficient expressiveness, the flexibility to inject
different levels of prior knowledge as well as the ability to transfer policies
between robots. However, implementing a system for end-to-end learning RMPflow
policies faces several computational challenges. In this work, we re-examine
the message passing algorithm of RMPflow and propose a more efficient alternate
algorithm, called RMP2, that uses modern automatic differentiation tools (such
as TensorFlow and PyTorch) to compute RMPflow policies. Our new design retains
the strengths of RMPflow while bringing in advantages from automatic
differentiation, including 1) easy programming interfaces to designing complex
transformations; 2) support of general directed acyclic graph (DAG)
transformation structures; 3) end-to-end differentiability for policy learning;
4) improved computational efficiency. Because of these features, RMP2 can be
treated as a structured policy class for efficient robot learning which is
suitable encoding domain knowledge. Our experiments show that using structured
policy class given by RMP2 can improve policy performance and safety in
reinforcement learning tasks for goal reaching in cluttered space.
Related papers
- Sparse Diffusion Policy: A Sparse, Reusable, and Flexible Policy for Robot Learning [61.294110816231886]
We introduce a sparse, reusable, and flexible policy, Sparse Diffusion Policy (SDP)
SDP selectively activates experts and skills, enabling efficient and task-specific learning without retraining the entire model.
Demos and codes can be found in https://forrest-110.io/sparse_diffusion_policy/.
arXiv Detail & Related papers (2024-07-01T17:59:56Z) - Riemannian Flow Matching Policy for Robot Motion Learning [5.724027955589408]
We introduce a novel model for learning and synthesizing robot visuomotor policies.
We show that RFMP provides smoother action trajectories with significantly lower inference times.
arXiv Detail & Related papers (2024-03-15T20:48:41Z) - Efficiently Learning Small Policies for Locomotion and Manipulation [12.340412143459869]
We leverage graph hyper networks to learn graph hyper policies trained with off-policy reinforcement learning.
We show that our method can be appended to any off-policy reinforcement learning algorithm.
We provide a method to select the best architecture, given a constraint on the number of parameters.
arXiv Detail & Related papers (2022-09-30T23:49:00Z) - XAI-N: Sensor-based Robot Navigation using Expert Policies and Decision
Trees [55.9643422180256]
We present a novel sensor-based learning navigation algorithm to compute a collision-free trajectory for a robot in dense and dynamic environments.
Our approach uses deep reinforcement learning-based expert policy that is trained using a sim2real paradigm.
We highlight the benefits of our algorithm in simulated environments and navigating a Clearpath Jackal robot among moving pedestrians.
arXiv Detail & Related papers (2021-04-22T01:33:10Z) - Composable Learning with Sparse Kernel Representations [110.19179439773578]
We present a reinforcement learning algorithm for learning sparse non-parametric controllers in a Reproducing Kernel Hilbert Space.
We improve the sample complexity of this approach by imposing a structure of the state-action function through a normalized advantage function.
We demonstrate the performance of this algorithm on learning obstacle-avoidance policies in multiple simulations of a robot equipped with a laser scanner while navigating in a 2D environment.
arXiv Detail & Related papers (2021-03-26T13:58:23Z) - Towards Coordinated Robot Motions: End-to-End Learning of Motion
Policies on Transform Trees [63.31965375413414]
We propose to solve multi-task problems through learning structured policies from human demonstrations.
Our structured policy is inspired by RMPflow, a framework for combining subtask policies on different spaces.
We derive an end-to-end learning objective function that is suitable for the multi-task problem.
arXiv Detail & Related papers (2020-12-24T22:46:22Z) - Learning High-Level Policies for Model Predictive Control [54.00297896763184]
Model Predictive Control (MPC) provides robust solutions to robot control tasks.
We propose a self-supervised learning algorithm for learning a neural network high-level policy.
We show that our approach can handle situations that are difficult for standard MPC.
arXiv Detail & Related papers (2020-07-20T17:12:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.