Deep Whole-Body Control: Learning a Unified Policy for Manipulation and
Locomotion
- URL: http://arxiv.org/abs/2210.10044v1
- Date: Tue, 18 Oct 2022 17:59:30 GMT
- Title: Deep Whole-Body Control: Learning a Unified Policy for Manipulation and
Locomotion
- Authors: Zipeng Fu, Xuxin Cheng, Deepak Pathak
- Abstract summary: An attached arm can significantly increase the applicability of legged robots to mobile manipulation tasks.
Standard hierarchical control pipeline for such legged manipulators is to decouple the controller into that of manipulation and locomotion.
We learn a unified policy for whole-body control of a legged manipulator using reinforcement learning.
- Score: 25.35885216505385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An attached arm can significantly increase the applicability of legged robots
to several mobile manipulation tasks that are not possible for the wheeled or
tracked counterparts. The standard hierarchical control pipeline for such
legged manipulators is to decouple the controller into that of manipulation and
locomotion. However, this is ineffective. It requires immense engineering to
support coordination between the arm and legs, and error can propagate across
modules causing non-smooth unnatural motions. It is also biological implausible
given evidence for strong motor synergies across limbs. In this work, we
propose to learn a unified policy for whole-body control of a legged
manipulator using reinforcement learning. We propose Regularized Online
Adaptation to bridge the Sim2Real gap for high-DoF control, and Advantage
Mixing exploiting the causal dependency in the action space to overcome local
minima during training the whole-body system. We also present a simple design
for a low-cost legged manipulator, and find that our unified policy can
demonstrate dynamic and agile behaviors across several task setups. Videos are
at https://maniploco.github.io
Related papers
- Learning Force Control for Legged Manipulation [18.894304288225385]
We propose a method for training RL policies for direct force control without requiring access to force sensing.
We showcase our method on a whole-body control platform of a quadruped robot with an arm.
We provide the first deployment of learned whole-body force control in legged manipulators, paving the way for more versatile and adaptable legged robots.
arXiv Detail & Related papers (2024-05-02T15:53:43Z) - Visual Whole-Body Control for Legged Loco-Manipulation [22.50054654508986]
We study the problem of mobile manipulation using legged robots equipped with an arm.
We propose a framework that can conduct the whole-body control autonomously with visual observations.
arXiv Detail & Related papers (2024-03-25T17:26:08Z) - DexDribbler: Learning Dexterous Soccer Manipulation via Dynamic Supervision [26.9579556496875]
Joint manipulation of moving objects and locomotion with legs, such as playing soccer, receive scant attention in the learning community.
We propose a feedback control block to compute the necessary body-level movement accurately and using the outputs as dynamic joint-level locomotion supervision.
We observe that our learning scheme can not only make the policy network converge faster but also enable soccer robots to perform sophisticated maneuvers.
arXiv Detail & Related papers (2024-03-21T11:16:28Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - Nonprehensile Planar Manipulation through Reinforcement Learning with
Multimodal Categorical Exploration [8.343657309038285]
Reinforcement Learning is a powerful framework for developing such robot controllers.
We propose a multimodal exploration approach through categorical distributions, which enables us to train planar pushing RL policies.
We show that the learned policies are robust to external disturbances and observation noise, and scale to tasks with multiple pushers.
arXiv Detail & Related papers (2023-08-04T16:55:00Z) - OSCAR: Data-Driven Operational Space Control for Adaptive and Robust
Robot Manipulation [50.59541802645156]
Operational Space Control (OSC) has been used as an effective task-space controller for manipulation.
We propose OSC for Adaptation and Robustness (OSCAR), a data-driven variant of OSC that compensates for modeling errors.
We evaluate our method on a variety of simulated manipulation problems, and find substantial improvements over an array of controller baselines.
arXiv Detail & Related papers (2021-10-02T01:21:38Z) - Reinforcement Learning for Robust Parameterized Locomotion Control of
Bipedal Robots [121.42930679076574]
We present a model-free reinforcement learning framework for training robust locomotion policies in simulation.
domain randomization is used to encourage the policies to learn behaviors that are robust across variations in system dynamics.
We demonstrate this on versatile walking behaviors such as tracking a target walking velocity, walking height, and turning yaw.
arXiv Detail & Related papers (2021-03-26T07:14:01Z) - Neural Dynamic Policies for End-to-End Sensorimotor Learning [51.24542903398335]
The current dominant paradigm in sensorimotor control, whether imitation or reinforcement learning, is to train policies directly in raw action spaces.
We propose Neural Dynamic Policies (NDPs) that make predictions in trajectory distribution space.
NDPs outperform the prior state-of-the-art in terms of either efficiency or performance across several robotic control tasks.
arXiv Detail & Related papers (2020-12-04T18:59:32Z) - Learning Quadrupedal Locomotion over Challenging Terrain [68.51539602703662]
Legged locomotion can dramatically expand the operational domains of robotics.
Conventional controllers for legged locomotion are based on elaborate state machines that explicitly trigger the execution of motion primitives and reflexes.
Here we present a radically robust controller for legged locomotion in challenging natural environments.
arXiv Detail & Related papers (2020-10-21T19:11:20Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.