Offline Adaptation of Quadruped Locomotion using Diffusion Models
- URL: http://arxiv.org/abs/2411.08832v1
- Date: Wed, 13 Nov 2024 18:12:15 GMT
- Title: Offline Adaptation of Quadruped Locomotion using Diffusion Models
- Authors: Reece O'Mahoney, Alexander L. Mitchell, Wanming Yu, Ingmar Posner, Ioannis Havoutis,
- Abstract summary: We present a diffusion-based approach to quadrupedal locomotion that simultaneously addresses the limitations of learning and interpolating between multiple skills.
We show that these capabilities are compatible with a multi-skill policy and can be applied with little modification and minimal compute overhead.
We verify the validity of our approach with hardware experiments on the ANYmal quadruped platform.
- Score: 59.882275766745295
- License:
- Abstract: We present a diffusion-based approach to quadrupedal locomotion that simultaneously addresses the limitations of learning and interpolating between multiple skills and of (modes) offline adapting to new locomotion behaviours after training. This is the first framework to apply classifier-free guided diffusion to quadruped locomotion and demonstrate its efficacy by extracting goal-conditioned behaviour from an originally unlabelled dataset. We show that these capabilities are compatible with a multi-skill policy and can be applied with little modification and minimal compute overhead, i.e., running entirely on the robots onboard CPU. We verify the validity of our approach with hardware experiments on the ANYmal quadruped platform.
Related papers
- Learning Quadruped Locomotion Using Differentiable Simulation [31.80380408663424]
Differentiable simulation promises fast convergence and stable training.
This work proposes a new differentiable simulation framework to overcome these challenges.
Our framework enables learning quadruped walking in simulation in minutes without parallelization.
arXiv Detail & Related papers (2024-03-21T22:18:59Z) - Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters [65.15700861265432]
We present a parameter-efficient continual learning framework to alleviate long-term forgetting in incremental learning with vision-language models.
Our approach involves the dynamic expansion of a pre-trained CLIP model, through the integration of Mixture-of-Experts (MoE) adapters.
To preserve the zero-shot recognition capability of vision-language models, we introduce a Distribution Discriminative Auto-Selector.
arXiv Detail & Related papers (2024-03-18T08:00:23Z) - Value function estimation using conditional diffusion models for control [62.27184818047923]
We propose a simple algorithm called Diffused Value Function (DVF)
It learns a joint multi-step model of the environment-robot interaction dynamics using a diffusion model.
We show how DVF can be used to efficiently capture the state visitation measure for multiple controllers.
arXiv Detail & Related papers (2023-06-09T18:40:55Z) - Learning a Shared Model for Motorized Prosthetic Joints to Predict
Ankle-Joint Motion [0.0]
We propose a learning-based shared model for predicting ankle-joint motion for different locomotion modes.
We show that the shared model is adequate for predicting the ankle angles and moments for different locomotion modes without explicitly classifying between the modes.
arXiv Detail & Related papers (2021-11-14T19:02:40Z) - STAR: Sparse Transformer-based Action Recognition [61.490243467748314]
This work proposes a novel skeleton-based human action recognition model with sparse attention on the spatial dimension and segmented linear attention on the temporal dimension of data.
Experiments show that our model can achieve comparable performance while utilizing much less trainable parameters and achieve high speed in training and inference.
arXiv Detail & Related papers (2021-07-15T02:53:11Z) - Learning Vision-Guided Quadrupedal Locomotion End-to-End with
Cross-Modal Transformers [14.509254362627576]
We propose to address quadrupedal locomotion tasks using Reinforcement Learning (RL)
We introduce LocoTransformer, an end-to-end RL method for quadrupedal locomotion.
arXiv Detail & Related papers (2021-07-08T17:41:55Z) - GLiDE: Generalizable Quadrupedal Locomotion in Diverse Environments with
a Centroidal Model [18.66472547798549]
We show how model-free reinforcement learning can be effectively used with a centroidal model to generate robust control policies for quadrupedal locomotion.
We show the potential of the method by demonstrating stepping-stone locomotion, two-legged in-place balance, balance beam locomotion, and sim-to-real transfer without further adaptations.
arXiv Detail & Related papers (2021-04-20T05:55:13Z) - Reinforcement Learning for Robust Parameterized Locomotion Control of
Bipedal Robots [121.42930679076574]
We present a model-free reinforcement learning framework for training robust locomotion policies in simulation.
domain randomization is used to encourage the policies to learn behaviors that are robust across variations in system dynamics.
We demonstrate this on versatile walking behaviors such as tracking a target walking velocity, walking height, and turning yaw.
arXiv Detail & Related papers (2021-03-26T07:14:01Z) - Efficient Learning of Control Policies for Robust Quadruped Bounding
using Pretrained Neural Networks [15.09037992110481]
Bounding is one of the important gaits in quadrupedal locomotion for negotiating obstacles.
The authors proposed an effective approach that can learn robust bounding gaits more efficiently.
The authors approach shows efficient computing and good locomotion results by the Jueying Mini quadrupedal robot bounding over uneven terrain.
arXiv Detail & Related papers (2020-11-01T08:06:46Z) - First Steps: Latent-Space Control with Semantic Constraints for
Quadruped Locomotion [73.37945453998134]
Traditional approaches to quadruped control employ simplified, hand-derived models.
This significantly reduces the capability of the robot since its effective kinematic range is curtailed.
In this work, these challenges are addressed by framing quadruped control as optimisation in a structured latent space.
A deep generative model captures a statistical representation of feasible joint configurations, whilst complex dynamic and terminal constraints are expressed via high-level, semantic indicators.
We validate the feasibility of locomotion trajectories optimised using our approach both in simulation and on a real-worldmal quadruped.
arXiv Detail & Related papers (2020-07-03T07:04:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.