ALLSTEPS: Curriculum-driven Learning of Stepping Stone Skills
- URL: http://arxiv.org/abs/2005.04323v2
- Date: Sun, 30 Aug 2020 00:45:00 GMT
- Title: ALLSTEPS: Curriculum-driven Learning of Stepping Stone Skills
- Authors: Zhaoming Xie, Hung Yu Ling, Nam Hee Kim, Michiel van de Panne
- Abstract summary: Finding good solutions to stepping-stone locomotion is a longstanding and fundamental challenge for animation and robotics.
We present fully learned solutions to this difficult problem using reinforcement learning.
Results are presented for a simulated human character, a realistic bipedal robot simulation and a monster character, in each case producing robust, plausible motions.
- Score: 8.406171678292964
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Humans are highly adept at walking in environments with foot placement
constraints, including stepping-stone scenarios where the footstep locations
are fully constrained. Finding good solutions to stepping-stone locomotion is a
longstanding and fundamental challenge for animation and robotics. We present
fully learned solutions to this difficult problem using reinforcement learning.
We demonstrate the importance of a curriculum for efficient learning and
evaluate four possible curriculum choices compared to a non-curriculum
baseline. Results are presented for a simulated human character, a realistic
bipedal robot simulation and a monster character, in each case producing
robust, plausible motions for challenging stepping stone sequences and
terrains.
Related papers
- Learning Humanoid Locomotion over Challenging Terrain [84.35038297708485]
We present a learning-based approach for blind humanoid locomotion capable of traversing challenging natural and man-made terrains.
Our model is first pre-trained on a dataset of flat-ground trajectories with sequence modeling, and then fine-tuned on uneven terrain using reinforcement learning.
We evaluate our model on a real humanoid robot across a variety of terrains, including rough, deformable, and sloped surfaces.
arXiv Detail & Related papers (2024-10-04T17:57:09Z) - VITAL: Visual Teleoperation to Enhance Robot Learning through Human-in-the-Loop Corrections [10.49712834719005]
We propose a low-cost visual teleoperation system for bimanual manipulation tasks, called VITAL.
Our approach leverages affordable hardware and visual processing techniques to collect demonstrations.
We enhance the generalizability and robustness of the learned policies by utilizing both real and simulated environments.
arXiv Detail & Related papers (2024-07-30T23:29:47Z) - Exciting Action: Investigating Efficient Exploration for Learning Musculoskeletal Humanoid Locomotion [16.63152794060493]
We demonstrate that adversarial imitation learning can address this issue by analyzing key problems and providing solutions.
We validate our methodology by learning walking and running gaits on a simulated humanoid model with 16 degrees of freedom and 92 Muscle-Tendon Units.
arXiv Detail & Related papers (2024-07-16T12:27:55Z) - Learning Robust, Agile, Natural Legged Locomotion Skills in the Wild [17.336553501547282]
We propose a new framework for learning robust, agile and natural legged locomotion skills over challenging terrain.
Empirical results on both simulation and real world of a quadruped robot demonstrate that our proposed algorithm enables robustly traversing challenging terrains.
arXiv Detail & Related papers (2023-04-21T11:09:23Z) - Learning Human-to-Robot Handovers from Point Clouds [63.18127198174958]
We propose the first framework to learn control policies for vision-based human-to-robot handovers.
We show significant performance gains over baselines on a simulation benchmark, sim-to-sim transfer and sim-to-real transfer.
arXiv Detail & Related papers (2023-03-30T17:58:36Z) - A Walk in the Park: Learning to Walk in 20 Minutes With Model-Free
Reinforcement Learning [86.06110576808824]
Deep reinforcement learning is a promising approach to learning policies in uncontrolled environments.
Recent advancements in machine learning algorithms and libraries combined with a carefully tuned robot controller lead to learning quadruped in only 20 minutes in the real world.
arXiv Detail & Related papers (2022-08-16T17:37:36Z) - What Matters in Learning from Offline Human Demonstrations for Robot
Manipulation [64.43440450794495]
We conduct an extensive study of six offline learning algorithms for robot manipulation.
Our study analyzes the most critical challenges when learning from offline human data.
We highlight opportunities for learning from human datasets.
arXiv Detail & Related papers (2021-08-06T20:48:30Z) - Reactive Long Horizon Task Execution via Visual Skill and Precondition
Models [59.76233967614774]
We describe an approach for sim-to-real training that can accomplish unseen robotic tasks using models learned in simulation to ground components of a simple task planner.
We show an increase in success rate from 91.6% to 98% in simulation and from 10% to 80% success rate in the real-world as compared with naive baselines.
arXiv Detail & Related papers (2020-11-17T15:24:01Z) - Guided Curriculum Learning for Walking Over Complex Terrain [2.4192504570921622]
We propose a 3-stage curriculum to train Deep Reinforcement Learning policies for bipedal walking over various challenging terrains.
In simulation experiments, we show that our approach is effective in learning walking policies, separate from each other, for five terrain types.
arXiv Detail & Related papers (2020-10-08T09:03:43Z) - Learning Agile Robotic Locomotion Skills by Imitating Animals [72.36395376558984]
Reproducing the diverse and agile locomotion skills of animals has been a longstanding challenge in robotics.
We present an imitation learning system that enables legged robots to learn agile locomotion skills by imitating real-world animals.
arXiv Detail & Related papers (2020-04-02T02:56:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.