Brain-Body-Task Co-Adaptation can Improve Autonomous Learning and Speed
of Bipedal Walking
- URL: http://arxiv.org/abs/2402.02387v1
- Date: Sun, 4 Feb 2024 07:57:52 GMT
- Title: Brain-Body-Task Co-Adaptation can Improve Autonomous Learning and Speed
of Bipedal Walking
- Authors: Dar\'io Urbina-Mel\'endez, Hesam Azadjou, Francisco J. Valero-Cuevas
- Abstract summary: Inspired by animals that co-adapt their brain and body to interact with the environment, we present a tendon-driven and over-actuated bipedal robot.
We show how continual physical adaptation can be driven by continual physical adaptation rooted in the backdrivable properties of the plant.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inspired by animals that co-adapt their brain and body to interact with the
environment, we present a tendon-driven and over-actuated (i.e., n joint, n+1
actuators) bipedal robot that (i) exploits its backdrivable mechanical
properties to manage body-environment interactions without explicit control,
and (ii) uses a simple 3-layer neural network to learn to walk after only 2
minutes of 'natural' motor babbling (i.e., an exploration strategy that is
compatible with leg and task dynamics; akin to childsplay). This brain-body
collaboration first learns to produce feet cyclical movements 'in air' and,
without further tuning, can produce locomotion when the biped is lowered to be
in slight contact with the ground. In contrast, training with 2 minutes of
'naive' motor babbling (i.e., an exploration strategy that ignores leg task
dynamics), does not produce consistent cyclical movements 'in air', and
produces erratic movements and no locomotion when in slight contact with the
ground. When further lowering the biped and making the desired leg trajectories
reach 1cm below ground (causing the desired-vs-obtained trajectories error to
be unavoidable), cyclical movements based on either natural or naive babbling
presented almost equally persistent trends, and locomotion emerged with naive
babbling. Therefore, we show how continual learning of walking in unforeseen
circumstances can be driven by continual physical adaptation rooted in the
backdrivable properties of the plant and enhanced by exploration strategies
that exploit plant dynamics. Our studies also demonstrate that the bio-inspired
codesign and co-adaptations of limbs and control strategies can produce
locomotion without explicit control of trajectory errors.
Related papers
- Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - Natural and Robust Walking using Reinforcement Learning without
Demonstrations in High-Dimensional Musculoskeletal Models [29.592874007260342]
Humans excel at robust bipedal walking in complex natural environments.
It is still not fully understood how the nervous system resolves the musculoskeletal redundancy to solve the multi-objective control problem.
arXiv Detail & Related papers (2023-09-06T13:20:31Z) - Persistent-Transient Duality: A Multi-mechanism Approach for Modeling
Human-Object Interaction [58.67761673662716]
Humans are highly adaptable, swiftly switching between different modes to handle different tasks, situations and contexts.
In Human-object interaction (HOI) activities, these modes can be attributed to two mechanisms: (1) the large-scale consistent plan for the whole activity and (2) the small-scale children interactive actions that start and end along the timeline.
This work proposes to model two concurrent mechanisms that jointly control human motion.
arXiv Detail & Related papers (2023-07-24T12:21:33Z) - Barkour: Benchmarking Animal-level Agility with Quadruped Robots [70.97471756305463]
We introduce the Barkour benchmark, an obstacle course to quantify agility for legged robots.
Inspired by dog agility competitions, it consists of diverse obstacles and a time based scoring mechanism.
We present two methods for tackling the benchmark.
arXiv Detail & Related papers (2023-05-24T02:49:43Z) - From Data-Fitting to Discovery: Interpreting the Neural Dynamics of
Motor Control through Reinforcement Learning [3.6159844753873087]
We study structured neural activity of a virtual robot performing legged locomotion.
We find that embodied agents trained to walk exhibit smooth dynamics that avoid tangling -- or opposing neural trajectories in neighboring neural space.
arXiv Detail & Related papers (2023-05-18T16:52:27Z) - An Adaptable Approach to Learn Realistic Legged Locomotion without
Examples [38.81854337592694]
This work proposes a generic approach for ensuring realism in locomotion by guiding the learning process with the spring-loaded inverted pendulum model as a reference.
We present experimental results showing that even in a model-free setup, the learned policies can generate realistic and energy-efficient locomotion gaits for a bipedal and a quadrupedal robot.
arXiv Detail & Related papers (2021-10-28T10:14:47Z) - Minimizing Energy Consumption Leads to the Emergence of Gaits in Legged
Robots [71.61319876928009]
We show that learning to minimize energy consumption plays a key role in the emergence of natural locomotion gaits at different speeds in real quadruped robots.
The emergent gaits are structured in ideal terrains and look similar to that of horses and sheep.
The same approach leads to unstructured gaits in rough terrains which is consistent with the findings in animal motor control.
arXiv Detail & Related papers (2021-10-25T17:59:58Z) - Reinforcement Learning for Robust Parameterized Locomotion Control of
Bipedal Robots [121.42930679076574]
We present a model-free reinforcement learning framework for training robust locomotion policies in simulation.
domain randomization is used to encourage the policies to learn behaviors that are robust across variations in system dynamics.
We demonstrate this on versatile walking behaviors such as tracking a target walking velocity, walking height, and turning yaw.
arXiv Detail & Related papers (2021-03-26T07:14:01Z) - Learning Quadrupedal Locomotion over Challenging Terrain [68.51539602703662]
Legged locomotion can dramatically expand the operational domains of robotics.
Conventional controllers for legged locomotion are based on elaborate state machines that explicitly trigger the execution of motion primitives and reflexes.
Here we present a radically robust controller for legged locomotion in challenging natural environments.
arXiv Detail & Related papers (2020-10-21T19:11:20Z) - Adaptive control for hindlimb locomotion in a simulated mouse through
temporal cerebellar learning [0.0]
Bio-inspired control system for adaptive locomotion of the musculoskeletal system of the mouse is presented.
A cerebellar-like module adapts the step time by using the double support interlimb asymmetry as a temporal teaching signal.
Results show adaptive locomotion behavior in the interlimb parameters similar to that seen in humans and mice.
arXiv Detail & Related papers (2020-02-07T14:25:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.