Continuous Versatile Jumping Using Learned Action Residuals
- URL: http://arxiv.org/abs/2304.08663v1
- Date: Mon, 17 Apr 2023 23:28:32 GMT
- Title: Continuous Versatile Jumping Using Learned Action Residuals
- Authors: Yuxiang Yang, Xiangyun Meng, Wenhao Yu, Tingnan Zhang, Jie Tan, Byron
Boots
- Abstract summary: We propose a hierarchical framework that combines optimal control and reinforcement learning to learn continuous jumping motions for quadrupedal robots.
The core of our framework is a stance controller, which combines a manually designed acceleration controller with a learned residual policy.
After training in simulation, our framework can be deployed directly to the real robot, and perform versatile, continuous jumping motions.
- Score: 35.996425893483796
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Jumping is essential for legged robots to traverse through difficult
terrains. In this work, we propose a hierarchical framework that combines
optimal control and reinforcement learning to learn continuous jumping motions
for quadrupedal robots. The core of our framework is a stance controller, which
combines a manually designed acceleration controller with a learned residual
policy. As the acceleration controller warm starts policy for efficient
training, the trained policy overcomes the limitation of the acceleration
controller and improves the jumping stability. In addition, a low-level
whole-body controller converts the body pose command from the stance controller
to motor commands. After training in simulation, our framework can be deployed
directly to the real robot, and perform versatile, continuous jumping motions,
including omni-directional jumps at up to 50cm high, 60cm forward, and
jump-turning at up to 90 degrees. Please visit our website for more results:
https://sites.google.com/view/learning-to-jump.
Related papers
- Impedance Matching: Enabling an RL-Based Running Jump in a Quadruped Robot [7.516046071926082]
We propose a new framework to mitigate the gap between simulated and real robots.
Our framework offers a structured guideline for parameter selection and the range for dynamics randomization in simulation.
Results are, to the best of our knowledge, one of the highest and longest running jumps demonstrated by an RL-based control policy in a real quadruped robot.
arXiv Detail & Related papers (2024-04-23T14:52:09Z) - Agile and versatile bipedal robot tracking control through reinforcement learning [12.831810518025309]
This paper proposes a versatile controller for bipedal robots.
It achieves ankle and body trajectory tracking across a wide range of gaits using a single small-scale neural network.
Highly flexible gait control can be achieved by combining minimal control units with high-level policy.
arXiv Detail & Related papers (2024-04-12T05:25:03Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - End-to-End Reinforcement Learning for Torque Based Variable Height
Hopping [5.34772724436823]
Legged locomotion is arguably the most suited and versatile mode to deal with natural or unstructured terrains.
In this paper, we present a end-to-end RL based torque controller that learns to implicitly detect the relevant jump phases.
We also extend a method for simulation to reality transfer of the learned controller to contact rich dynamic tasks, resulting in successful deployment on the robot.
arXiv Detail & Related papers (2023-07-31T13:51:29Z) - Barkour: Benchmarking Animal-level Agility with Quadruped Robots [70.97471756305463]
We introduce the Barkour benchmark, an obstacle course to quantify agility for legged robots.
Inspired by dog agility competitions, it consists of diverse obstacles and a time based scoring mechanism.
We present two methods for tackling the benchmark.
arXiv Detail & Related papers (2023-05-24T02:49:43Z) - Learning and Adapting Agile Locomotion Skills by Transferring Experience [71.8926510772552]
We propose a framework for training complex robotic skills by transferring experience from existing controllers to jumpstart learning new tasks.
We show that our method enables learning complex agile jumping behaviors, navigating to goal locations while walking on hind legs, and adapting to new environments.
arXiv Detail & Related papers (2023-04-19T17:37:54Z) - Robust and Versatile Bipedal Jumping Control through Reinforcement
Learning [141.56016556936865]
This work aims to push the limits of agility for bipedal robots by enabling a torque-controlled bipedal robot to perform robust and versatile dynamic jumps in the real world.
We present a reinforcement learning framework for training a robot to accomplish a large variety of jumping tasks, such as jumping to different locations and directions.
We develop a new policy structure that encodes the robot's long-term input/output (I/O) history while also providing direct access to a short-term I/O history.
arXiv Detail & Related papers (2023-02-19T01:06:09Z) - Fast and Efficient Locomotion via Learned Gait Transitions [35.86279693549959]
We focus on the problem of developing efficient controllers for quadrupedal robots.
We devise a hierarchical learning framework, in which distinctive locomotion gaits and natural gait transitions emerge automatically.
We show that the learned hierarchical controller consumes much less energy across a wide range of locomotion speed than baseline controllers.
arXiv Detail & Related papers (2021-04-09T23:53:28Z) - Reinforcement Learning for Robust Parameterized Locomotion Control of
Bipedal Robots [121.42930679076574]
We present a model-free reinforcement learning framework for training robust locomotion policies in simulation.
domain randomization is used to encourage the policies to learn behaviors that are robust across variations in system dynamics.
We demonstrate this on versatile walking behaviors such as tracking a target walking velocity, walking height, and turning yaw.
arXiv Detail & Related papers (2021-03-26T07:14:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.