Co-learning Planning and Control Policies Constrained by Differentiable
Logic Specifications
- URL: http://arxiv.org/abs/2303.01346v3
- Date: Mon, 2 Oct 2023 03:45:15 GMT
- Title: Co-learning Planning and Control Policies Constrained by Differentiable
Logic Specifications
- Authors: Zikang Xiong, Daniel Lawson, Joe Eappen, Ahmed H. Qureshi, Suresh
Jagannathan
- Abstract summary: This paper presents a novel reinforcement learning approach to solving high-dimensional robot navigation tasks.
We train high-quality policies with much fewer samples compared to existing reinforcement learning algorithms.
Our approach also demonstrates capabilities for high-dimensional control and avoiding suboptimal policies via policy alignment.
- Score: 4.12484724941528
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Synthesizing planning and control policies in robotics is a fundamental task,
further complicated by factors such as complex logic specifications and
high-dimensional robot dynamics. This paper presents a novel reinforcement
learning approach to solving high-dimensional robot navigation tasks with
complex logic specifications by co-learning planning and control policies.
Notably, this approach significantly reduces the sample complexity in training,
allowing us to train high-quality policies with much fewer samples compared to
existing reinforcement learning algorithms. In addition, our methodology
streamlines complex specification extraction from map images and enables the
efficient generation of long-horizon robot motion paths across different map
layouts. Moreover, our approach also demonstrates capabilities for
high-dimensional control and avoiding suboptimal policies via policy alignment.
The efficacy of our approach is demonstrated through experiments involving
simulated high-dimensional quadruped robot dynamics and a real-world
differential drive robot (TurtleBot3) under different types of task
specifications.
Related papers
- Bridging the gap between Learning-to-plan, Motion Primitives and Safe Reinforcement Learning [20.158498233576143]
Trajectory planning under kinodynamic constraints is fundamental for advanced robotics applications.
Recent advances in kinodynamic planning demonstrate that learning-to-plan techniques can generate complex motions under intricate constraints.
This paper addresses this limitation by combining learning-to-plan methods with reinforcement learning, resulting in a novel integration of black-box learning of motion primitives and optimization.
arXiv Detail & Related papers (2024-08-26T07:44:53Z) - Mission-driven Exploration for Accelerated Deep Reinforcement Learning
with Temporal Logic Task Specifications [11.812602599752294]
We consider robots with unknown dynamics operating in environments with unknown structure.
Our goal is to synthesize a control policy that maximizes the probability of satisfying an automaton-encoded task.
We propose a novel DRL algorithm, which has the capability to learn control policies at a notably faster rate compared to similar methods.
arXiv Detail & Related papers (2023-11-28T18:59:58Z) - Nonprehensile Planar Manipulation through Reinforcement Learning with
Multimodal Categorical Exploration [8.343657309038285]
Reinforcement Learning is a powerful framework for developing such robot controllers.
We propose a multimodal exploration approach through categorical distributions, which enables us to train planar pushing RL policies.
We show that the learned policies are robust to external disturbances and observation noise, and scale to tasks with multiple pushers.
arXiv Detail & Related papers (2023-08-04T16:55:00Z) - Multi-Robot Deep Reinforcement Learning for Mobile Navigation [82.62621210336881]
We propose a deep reinforcement learning algorithm with hierarchically integrated models (HInt)
At training time, HInt learns separate perception and dynamics models, and at test time, HInt integrates the two models in a hierarchical manner and plans actions with the integrated model.
Our mobile navigation experiments show that HInt outperforms conventional hierarchical policies and single-source approaches.
arXiv Detail & Related papers (2021-06-24T19:07:40Z) - Simultaneous Navigation and Construction Benchmarking Environments [73.0706832393065]
We need intelligent robots for mobile construction, the process of navigating in an environment and modifying its structure according to a geometric design.
In this task, a major robot vision and learning challenge is how to exactly achieve the design without GPS.
We benchmark the performance of a handcrafted policy with basic localization and planning, and state-of-the-art deep reinforcement learning methods.
arXiv Detail & Related papers (2021-03-31T00:05:54Z) - Composable Learning with Sparse Kernel Representations [110.19179439773578]
We present a reinforcement learning algorithm for learning sparse non-parametric controllers in a Reproducing Kernel Hilbert Space.
We improve the sample complexity of this approach by imposing a structure of the state-action function through a normalized advantage function.
We demonstrate the performance of this algorithm on learning obstacle-avoidance policies in multiple simulations of a robot equipped with a laser scanner while navigating in a 2D environment.
arXiv Detail & Related papers (2021-03-26T13:58:23Z) - Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic
Platforms [60.59764170868101]
Reinforcement learning methods can achieve significant performance but require a large amount of training data collected on the same robotic platform.
We formulate it as a few-shot meta-learning problem where the goal is to find a model that captures the common structure shared across different robotic platforms.
We experimentally evaluate our framework on a simulated reaching and a real-robot picking task using 400 simulated robots.
arXiv Detail & Related papers (2021-03-05T14:16:20Z) - Towards Coordinated Robot Motions: End-to-End Learning of Motion
Policies on Transform Trees [63.31965375413414]
We propose to solve multi-task problems through learning structured policies from human demonstrations.
Our structured policy is inspired by RMPflow, a framework for combining subtask policies on different spaces.
We derive an end-to-end learning objective function that is suitable for the multi-task problem.
arXiv Detail & Related papers (2020-12-24T22:46:22Z) - Learning High-Level Policies for Model Predictive Control [54.00297896763184]
Model Predictive Control (MPC) provides robust solutions to robot control tasks.
We propose a self-supervised learning algorithm for learning a neural network high-level policy.
We show that our approach can handle situations that are difficult for standard MPC.
arXiv Detail & Related papers (2020-07-20T17:12:34Z) - Learning compositional models of robot skills for task and motion
planning [39.36562555272779]
We learn to use sensorimotor primitives to solve complex long-horizon manipulation problems.
We use state-of-the-art methods for active learning and sampling.
We evaluate our approach both in simulation and in the real world through measuring the quality of the selected primitive actions.
arXiv Detail & Related papers (2020-06-08T20:45:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.