Next Steps: Learning a Disentangled Gait Representation for Versatile
Quadruped Locomotion
- URL: http://arxiv.org/abs/2112.04809v1
- Date: Thu, 9 Dec 2021 10:02:02 GMT
- Title: Next Steps: Learning a Disentangled Gait Representation for Versatile
Quadruped Locomotion
- Authors: Alexander L. Mitchell, Wolfgang Merkt, Mathieu Geisert, Siddhant
Gangapurwala, Martin Engelcke, Oiwi Parker Jones, Ioannis Havoutis, and
Ingmar Posner
- Abstract summary: Current planners are unable to vary key gait parameters continuously while the robot is in motion.
In this work we address this limitation by learning a latent space capturing the key stance phases constituting a particular gait.
We demonstrate that specific properties of the drive signal map directly to gait parameters such as cadence, foot step height and full stance duration.
- Score: 69.87112582900363
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Quadruped locomotion is rapidly maturing to a degree where robots now
routinely traverse a variety of unstructured terrains. However, while gaits can
be varied typically by selecting from a range of pre-computed styles, current
planners are unable to vary key gait parameters continuously while the robot is
in motion. The synthesis, on-the-fly, of gaits with unexpected operational
characteristics or even the blending of dynamic manoeuvres lies beyond the
capabilities of the current state-of-the-art. In this work we address this
limitation by learning a latent space capturing the key stance phases
constituting a particular gait. This is achieved via a generative model trained
on a single trot style, which encourages disentanglement such that application
of a drive signal to a single dimension of the latent state induces holistic
plans synthesising a continuous variety of trot styles. We demonstrate that
specific properties of the drive signal map directly to gait parameters such as
cadence, foot step height and full stance duration. Due to the nature of our
approach these synthesised gaits are continuously variable online during robot
operation and robustly capture a richness of movement significantly exceeding
the relatively narrow behaviour seen during training. In addition, the use of a
generative model facilitates the detection and mitigation of disturbances to
provide a versatile and robust planning framework. We evaluate our approach on
a real ANYmal quadruped robot and demonstrate that our method achieves a
continuous blend of dynamic trot styles whilst being robust and reactive to
external perturbations.
Related papers
- Phase-Amplitude Reduction-Based Imitation Learning [0.9668407688201356]
We propose the use of the phase-amplitude reduction method to construct an imitation learning framework.
Unlike previous dynamical system-based imitation learning approaches, our proposed method allows the robot to imitate a limit cycle trajectory.
Our findings confirm that our proposed method can more accurately generate transient movements to converge on a target periodic attractor.
arXiv Detail & Related papers (2024-06-06T04:19:55Z) - Persistent-Transient Duality: A Multi-mechanism Approach for Modeling
Human-Object Interaction [58.67761673662716]
Humans are highly adaptable, swiftly switching between different modes to handle different tasks, situations and contexts.
In Human-object interaction (HOI) activities, these modes can be attributed to two mechanisms: (1) the large-scale consistent plan for the whole activity and (2) the small-scale children interactive actions that start and end along the timeline.
This work proposes to model two concurrent mechanisms that jointly control human motion.
arXiv Detail & Related papers (2023-07-24T12:21:33Z) - Simultaneous Contact-Rich Grasping and Locomotion via Distributed
Optimization Enabling Free-Climbing for Multi-Limbed Robots [60.06216976204385]
We present an efficient motion planning framework for simultaneously solving locomotion, grasping, and contact problems.
We demonstrate our proposed framework in the hardware experiments, showing that the multi-limbed robot is able to realize various motions including free-climbing at a slope angle 45deg with a much shorter planning time.
arXiv Detail & Related papers (2022-07-04T13:52:10Z) - Weakly-supervised Action Transition Learning for Stochastic Human Motion
Prediction [81.94175022575966]
We introduce the task of action-driven human motion prediction.
It aims to predict multiple plausible future motions given a sequence of action labels and a short motion history.
arXiv Detail & Related papers (2022-05-31T08:38:07Z) - VAE-Loco: Versatile Quadruped Locomotion by Learning a Disentangled Gait
Representation [78.92147339883137]
We show that it is pivotal in increasing controller robustness by learning a latent space capturing the key stance phases constituting a particular gait.
We demonstrate that specific properties of the drive signal map directly to gait parameters such as cadence, footstep height and full stance duration.
The use of a generative model facilitates the detection and mitigation of disturbances to provide a versatile and robust planning framework.
arXiv Detail & Related papers (2022-05-02T19:49:53Z) - Learning to Jump from Pixels [23.17535989519855]
We present Depth-based Impulse Control (DIC), a method for synthesizing highly agile visually-guided behaviors.
DIC affords the flexibility of model-free learning but regularizes behavior through explicit model-based optimization of ground reaction forces.
We evaluate the proposed method both in simulation and in the real world.
arXiv Detail & Related papers (2021-10-28T17:53:06Z) - An Adaptable Approach to Learn Realistic Legged Locomotion without
Examples [38.81854337592694]
This work proposes a generic approach for ensuring realism in locomotion by guiding the learning process with the spring-loaded inverted pendulum model as a reference.
We present experimental results showing that even in a model-free setup, the learned policies can generate realistic and energy-efficient locomotion gaits for a bipedal and a quadrupedal robot.
arXiv Detail & Related papers (2021-10-28T10:14:47Z) - First Steps: Latent-Space Control with Semantic Constraints for
Quadruped Locomotion [73.37945453998134]
Traditional approaches to quadruped control employ simplified, hand-derived models.
This significantly reduces the capability of the robot since its effective kinematic range is curtailed.
In this work, these challenges are addressed by framing quadruped control as optimisation in a structured latent space.
A deep generative model captures a statistical representation of feasible joint configurations, whilst complex dynamic and terminal constraints are expressed via high-level, semantic indicators.
We validate the feasibility of locomotion trajectories optimised using our approach both in simulation and on a real-worldmal quadruped.
arXiv Detail & Related papers (2020-07-03T07:04:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.