Reinforcement Learning with Prior Policy Guidance for Motion Planning of
Dual-Arm Free-Floating Space Robot
- URL: http://arxiv.org/abs/2209.01434v1
- Date: Sat, 3 Sep 2022 14:20:17 GMT
- Title: Reinforcement Learning with Prior Policy Guidance for Motion Planning of
Dual-Arm Free-Floating Space Robot
- Authors: Yuxue Cao, Shengjie Wang, Xiang Zheng, Wenke Ma, Xinru Xie, Lei Liu
- Abstract summary: We propose a novel algorithm, Efficient, to facilitate RL-based methods to improve planning accuracy efficiently.
Our core contributions are constructing a mixed policy with prior knowledge guidance and introducing infinite norm to build a more reasonable reward function.
- Score: 11.272278713797537
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning methods as a promising technique have achieved
superior results in the motion planning of free-floating space robots. However,
due to the increase in planning dimension and the intensification of system
dynamics coupling, the motion planning of dual-arm free-floating space robots
remains an open challenge. In particular, the current study cannot handle the
task of capturing a non-cooperative object due to the lack of the pose
constraint of the end-effectors. To address the problem, we propose a novel
algorithm, EfficientLPT, to facilitate RL-based methods to improve planning
accuracy efficiently. Our core contributions are constructing a mixed policy
with prior knowledge guidance and introducing infinite norm to build a more
reasonable reward function. Furthermore, our method successfully captures a
rotating object with different spinning speeds.
Related papers
- Trajectory Manifold Optimization for Fast and Adaptive Kinodynamic Motion Planning [5.982922468400902]
Fast kinodynamic motion planning is crucial for systems to adapt to dynamically changing environments.
We propose a novel neural network model, it Differentiable Motion Manifold Primitives (DMMP), along with a practical training strategy.
Experiments with a 7-DoF robot arm tasked with dynamic throwing to arbitrary target positions demonstrate that our method surpasses existing approaches in planning speed, task success, and constraint satisfaction.
arXiv Detail & Related papers (2024-10-16T03:29:33Z) - Bridging the gap between Learning-to-plan, Motion Primitives and Safe Reinforcement Learning [20.158498233576143]
Trajectory planning under kinodynamic constraints is fundamental for advanced robotics applications.
Recent advances in kinodynamic planning demonstrate that learning-to-plan techniques can generate complex motions under intricate constraints.
This paper addresses this limitation by combining learning-to-plan methods with reinforcement learning, resulting in a novel integration of black-box learning of motion primitives and optimization.
arXiv Detail & Related papers (2024-08-26T07:44:53Z) - Potential Based Diffusion Motion Planning [73.593988351275]
We propose a new approach towards learning potential based motion planning.
We train a neural network to capture and learn an easily optimizable potentials over motion planning trajectories.
We demonstrate its inherent composability, enabling us to generalize to a multitude of different motion constraints.
arXiv Detail & Related papers (2024-07-08T17:48:39Z) - Safe multi-agent motion planning under uncertainty for drones using
filtered reinforcement learning [6.783774261623415]
We present a tractable motion planner that builds upon the strengths of reinforcement learning and constrained-control-based trajectory planning.
The proposed approach yields a safe, real-time implementable, multi-agent motion planner that is simpler to train than methods based solely on learning.
arXiv Detail & Related papers (2023-10-31T18:09:26Z) - Leveraging Sequentiality in Reinforcement Learning from a Single
Demonstration [68.94506047556412]
We propose to leverage a sequential bias to learn control policies for complex robotic tasks using a single demonstration.
We show that DCIL-II can solve with unprecedented sample efficiency some challenging simulated tasks such as humanoid locomotion and stand-up.
arXiv Detail & Related papers (2022-11-09T10:28:40Z) - Reaching Through Latent Space: From Joint Statistics to Path Planning in
Manipulation [26.38185646091712]
We present a novel approach to path planning for robotic manipulators.
Paths are produced via iterative optimisation in the latent space of a generative model of robot poses.
Our models are trained in a task-agnostic manner on randomly sampled robot poses.
arXiv Detail & Related papers (2022-10-21T07:25:21Z) - A Learning System for Motion Planning of Free-Float Dual-Arm Space
Manipulator towards Non-Cooperative Object [13.289739243378245]
We propose a learning system for motion planning of free-float dual-arm space manipulator (FFDASM) towards non-cooperative objects.
Module I realizes the multi-target trajectory planning for two end-effectors within a large target space.
Module II takes as input the point clouds of the non-cooperative object to estimate the motional property, and then can predict the position of target points on a non-cooperative object.
arXiv Detail & Related papers (2022-07-06T06:22:34Z) - Simultaneous Contact-Rich Grasping and Locomotion via Distributed
Optimization Enabling Free-Climbing for Multi-Limbed Robots [60.06216976204385]
We present an efficient motion planning framework for simultaneously solving locomotion, grasping, and contact problems.
We demonstrate our proposed framework in the hardware experiments, showing that the multi-limbed robot is able to realize various motions including free-climbing at a slope angle 45deg with a much shorter planning time.
arXiv Detail & Related papers (2022-07-04T13:52:10Z) - Reinforcement Learning for Low-Thrust Trajectory Design of
Interplanetary Missions [77.34726150561087]
This paper investigates the use of reinforcement learning for the robust design of interplanetary trajectories in presence of severe disturbances.
An open-source implementation of the state-of-the-art algorithm Proximal Policy Optimization is adopted.
The resulting Guidance and Control Network provides both a robust nominal trajectory and the associated closed-loop guidance law.
arXiv Detail & Related papers (2020-08-19T15:22:15Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z) - Reinforcement Learning with Fast Stabilization in Linear Dynamical
Systems [91.43582419264763]
We study model-based reinforcement learning (RL) in unknown stabilizable linear dynamical systems.
We propose an algorithm that certifies fast stabilization of the underlying system by effectively exploring the environment.
We show that the proposed algorithm attains $tildemathcalO(sqrtT)$ regret after $T$ time steps of agent-environment interaction.
arXiv Detail & Related papers (2020-07-23T23:06:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.