Learning Torque Control for Quadrupedal Locomotion
- URL: http://arxiv.org/abs/2203.05194v1
- Date: Thu, 10 Mar 2022 07:09:05 GMT
- Title: Learning Torque Control for Quadrupedal Locomotion
- Authors: Shuxiao Chen, Bike Zhang, Mark W. Mueller, Akshara Rai and Koushil
Sreenath
- Abstract summary: This paper introduces a learning torque control framework for quadrupedal locomotion.
It trains an RL policy that directly predicts joint torques at a high frequency, thus circumventing the use of PD controllers.
To our knowledge, this is the first attempt of learning torque control for quadrupedal locomotion with an end-to-end single neural network.
- Score: 22.415419916292187
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning (RL) is a promising tool for developing controllers
for quadrupedal locomotion. The design of most learning-based locomotion
controllers adopts the joint position-based paradigm, wherein a low-frequency
RL policy outputs target joint positions that are then tracked by a
high-frequency proportional-derivative (PD) controller that outputs joint
torques. However, the low frequency of such a policy hinders the advancement of
highly dynamic locomotion behaviors. Moreover, determining the PD gains for
optimal tracking performance is laborious and dependent on the task at hand. In
this paper, we introduce a learning torque control framework for quadrupedal
locomotion, which trains an RL policy that directly predicts joint torques at a
high frequency, thus circumventing the use of PD controllers. We validate the
proposed framework with extensive experiments where the robot is able to both
traverse various terrains and resist external pushes, given user-specified
commands. To our knowledge, this is the first attempt of learning torque
control for quadrupedal locomotion with an end-to-end single neural network
that has led to successful real-world experiments among recent research on
learning-based quadrupedal locomotion which is mostly position-based.
Related papers
- Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - TLControl: Trajectory and Language Control for Human Motion Synthesis [68.09806223962323]
We present TLControl, a novel method for realistic human motion synthesis.
It incorporates both low-level Trajectory and high-level Language semantics controls.
It is practical for interactive and high-quality animation generation.
arXiv Detail & Related papers (2023-11-28T18:54:16Z) - End-to-End Reinforcement Learning for Torque Based Variable Height
Hopping [5.34772724436823]
Legged locomotion is arguably the most suited and versatile mode to deal with natural or unstructured terrains.
In this paper, we present a end-to-end RL based torque controller that learns to implicitly detect the relevant jump phases.
We also extend a method for simulation to reality transfer of the learned controller to contact rich dynamic tasks, resulting in successful deployment on the robot.
arXiv Detail & Related papers (2023-07-31T13:51:29Z) - Learning Low-Frequency Motion Control for Robust and Dynamic Robot
Locomotion [10.838285018473725]
We demonstrate robust and dynamic locomotion with a learned motion controller executing at as low as 8 Hz on a real ANYmal C quadruped.
The robot is able to robustly and repeatably achieve a high heading velocity of 1.5 m/s, traverse uneven terrain, and resist unexpected external perturbations.
arXiv Detail & Related papers (2022-09-29T15:55:33Z) - Learning a Single Near-hover Position Controller for Vastly Different
Quadcopters [56.37274861303324]
This paper proposes an adaptive near-hover position controller for quadcopters.
It can be deployed to quadcopters of very different mass, size and motor constants.
It also shows rapid adaptation to unknown disturbances during runtime.
arXiv Detail & Related papers (2022-09-19T17:55:05Z) - VAE-Loco: Versatile Quadruped Locomotion by Learning a Disentangled Gait
Representation [78.92147339883137]
We show that it is pivotal in increasing controller robustness by learning a latent space capturing the key stance phases constituting a particular gait.
We demonstrate that specific properties of the drive signal map directly to gait parameters such as cadence, footstep height and full stance duration.
The use of a generative model facilitates the detection and mitigation of disturbances to provide a versatile and robust planning framework.
arXiv Detail & Related papers (2022-05-02T19:49:53Z) - GLiDE: Generalizable Quadrupedal Locomotion in Diverse Environments with
a Centroidal Model [18.66472547798549]
We show how model-free reinforcement learning can be effectively used with a centroidal model to generate robust control policies for quadrupedal locomotion.
We show the potential of the method by demonstrating stepping-stone locomotion, two-legged in-place balance, balance beam locomotion, and sim-to-real transfer without further adaptations.
arXiv Detail & Related papers (2021-04-20T05:55:13Z) - Reinforcement Learning for Robust Parameterized Locomotion Control of
Bipedal Robots [121.42930679076574]
We present a model-free reinforcement learning framework for training robust locomotion policies in simulation.
domain randomization is used to encourage the policies to learn behaviors that are robust across variations in system dynamics.
We demonstrate this on versatile walking behaviors such as tracking a target walking velocity, walking height, and turning yaw.
arXiv Detail & Related papers (2021-03-26T07:14:01Z) - Neural Dynamic Policies for End-to-End Sensorimotor Learning [51.24542903398335]
The current dominant paradigm in sensorimotor control, whether imitation or reinforcement learning, is to train policies directly in raw action spaces.
We propose Neural Dynamic Policies (NDPs) that make predictions in trajectory distribution space.
NDPs outperform the prior state-of-the-art in terms of either efficiency or performance across several robotic control tasks.
arXiv Detail & Related papers (2020-12-04T18:59:32Z) - Efficient Learning of Control Policies for Robust Quadruped Bounding
using Pretrained Neural Networks [15.09037992110481]
Bounding is one of the important gaits in quadrupedal locomotion for negotiating obstacles.
The authors proposed an effective approach that can learn robust bounding gaits more efficiently.
The authors approach shows efficient computing and good locomotion results by the Jueying Mini quadrupedal robot bounding over uneven terrain.
arXiv Detail & Related papers (2020-11-01T08:06:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.