ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically
Simulated Characters
- URL: http://arxiv.org/abs/2205.01906v2
- Date: Thu, 5 May 2022 17:25:14 GMT
- Title: ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically
Simulated Characters
- Authors: Xue Bin Peng, Yunrong Guo, Lina Halper, Sergey Levine, Sanja Fidler
- Abstract summary: General-purpose motor skills enable humans to perform complex tasks.
These skills also provide powerful priors for guiding their behaviors when learning new tasks.
We present a framework for learning versatile and reusable skill embeddings for physically simulated characters.
- Score: 123.88692739360457
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The incredible feats of athleticism demonstrated by humans are made possible
in part by a vast repertoire of general-purpose motor skills, acquired through
years of practice and experience. These skills not only enable humans to
perform complex tasks, but also provide powerful priors for guiding their
behaviors when learning new tasks. This is in stark contrast to what is common
practice in physics-based character animation, where control policies are most
typically trained from scratch for each task. In this work, we present a
large-scale data-driven framework for learning versatile and reusable skill
embeddings for physically simulated characters. Our approach combines
techniques from adversarial imitation learning and unsupervised reinforcement
learning to develop skill embeddings that produce life-like behaviors, while
also providing an easy to control representation for use on new downstream
tasks. Our models can be trained using large datasets of unstructured motion
clips, without requiring any task-specific annotation or segmentation of the
motion data. By leveraging a massively parallel GPU-based simulator, we are
able to train skill embeddings using over a decade of simulated experiences,
enabling our model to learn a rich and versatile repertoire of skills. We show
that a single pre-trained model can be effectively applied to perform a diverse
set of new tasks. Our system also allows users to specify tasks through simple
reward functions, and the skill embedding then enables the character to
automatically synthesize complex and naturalistic strategies in order to
achieve the task objectives.
Related papers
- HYPERmotion: Learning Hybrid Behavior Planning for Autonomous Loco-manipulation [7.01404330241523]
HYPERmotion is a framework that learns, selects and plans behaviors based on tasks in different scenarios.
We combine reinforcement learning with whole-body optimization to generate motion for 38 actuated joints.
Experiments in simulation and real-world show that learned motions can efficiently adapt to new tasks.
arXiv Detail & Related papers (2024-06-20T18:21:24Z) - C$\cdot$ASE: Learning Conditional Adversarial Skill Embeddings for
Physics-based Characters [49.83342243500835]
We present C$cdot$ASE, an efficient framework that learns conditional Adversarial Skill Embeddings for physics-based characters.
C$cdot$ASE divides the heterogeneous skill motions into distinct subsets containing homogeneous samples for training a low-level conditional model.
The skill-conditioned imitation learning naturally offers explicit control over the character's skills after training.
arXiv Detail & Related papers (2023-09-20T14:34:45Z) - Hierarchical Few-Shot Imitation with Skill Transition Models [66.81252581083199]
Few-shot Imitation with Skill Transition Models (FIST) is an algorithm that extracts skills from offline data and utilizes them to generalize to unseen tasks.
We show that FIST is capable of generalizing to new tasks and substantially outperforms prior baselines in navigation experiments.
arXiv Detail & Related papers (2021-07-19T15:56:01Z) - MT-Opt: Continuous Multi-Task Robotic Reinforcement Learning at Scale [103.7609761511652]
We show how a large-scale collective robotic learning system can acquire a repertoire of behaviors simultaneously.
New tasks can be continuously instantiated from previously learned tasks.
We train and evaluate our system on a set of 12 real-world tasks with data collected from 7 robots.
arXiv Detail & Related papers (2021-04-16T16:38:02Z) - Emergent Real-World Robotic Skills via Unsupervised Off-Policy
Reinforcement Learning [81.12201426668894]
We develop efficient reinforcement learning methods that acquire diverse skills without any reward function, and then repurpose these skills for downstream tasks.
We show that our proposed algorithm provides substantial improvement in learning efficiency, making reward-free real-world training feasible.
We also demonstrate that the learned skills can be composed using model predictive control for goal-oriented navigation, without any additional training.
arXiv Detail & Related papers (2020-04-27T17:38:53Z) - Learning Agile Robotic Locomotion Skills by Imitating Animals [72.36395376558984]
Reproducing the diverse and agile locomotion skills of animals has been a longstanding challenge in robotics.
We present an imitation learning system that enables legged robots to learn agile locomotion skills by imitating real-world animals.
arXiv Detail & Related papers (2020-04-02T02:56:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.