Hierarchical generative modelling for autonomous robots
- URL: http://arxiv.org/abs/2308.07775v1
- Date: Tue, 15 Aug 2023 13:51:03 GMT
- Title: Hierarchical generative modelling for autonomous robots
- Authors: Kai Yuan, Noor Sajid, Karl Friston, Zhibin Li
- Abstract summary: We show how a humanoid robot can autonomously complete a complex task that requires a holistic use of locomotion, manipulation, and grasping.
Specifically, we demonstrate the ability of a humanoid robot that can retrieve and transport a box, open and walk through a door to reach the destination, approach and kick a football, while showing robust performance in presence of body damage and ground irregularities.
- Score: 8.023920215148486
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Humans can produce complex whole-body motions when interacting with their
surroundings, by planning, executing and combining individual limb movements.
We investigated this fundamental aspect of motor control in the setting of
autonomous robotic operations. We approach this problem by hierarchical
generative modelling equipped with multi-level planning-for autonomous task
completion-that mimics the deep temporal architecture of human motor control.
Here, temporal depth refers to the nested time scales at which successive
levels of a forward or generative model unfold, for example, delivering an
object requires a global plan to contextualise the fast coordination of
multiple local movements of limbs. This separation of temporal scales also
motivates robotics and control. Specifically, to achieve versatile sensorimotor
control, it is advantageous to hierarchically structure the planning and
low-level motor control of individual limbs. We use numerical and physical
simulation to conduct experiments and to establish the efficacy of this
formulation. Using a hierarchical generative model, we show how a humanoid
robot can autonomously complete a complex task that necessitates a holistic use
of locomotion, manipulation, and grasping. Specifically, we demonstrate the
ability of a humanoid robot that can retrieve and transport a box, open and
walk through a door to reach the destination, approach and kick a football,
while showing robust performance in presence of body damage and ground
irregularities. Our findings demonstrated the effectiveness of using
human-inspired motor control algorithms, and our method provides a viable
hierarchical architecture for the autonomous completion of challenging
goal-directed tasks.
Related papers
- Grounding Language Models in Autonomous Loco-manipulation Tasks [3.8363685417355557]
We propose a novel framework that learns, selects, and plans behaviors based on tasks in different scenarios.
We leverage the planning and reasoning features of the large language model (LLM), constructing a hierarchical task graph.
Experiments in simulation and real-world using the CENTAURO robot show that the language model based planner can efficiently adapt to new loco-manipulation tasks.
arXiv Detail & Related papers (2024-09-02T15:27:48Z) - Commonsense Reasoning for Legged Robot Adaptation with Vision-Language Models [81.55156507635286]
Legged robots are physically capable of navigating a diverse variety of environments and overcoming a wide range of obstructions.
Current learning methods often struggle with generalization to the long tail of unexpected situations without heavy human supervision.
We propose a system, VLM-Predictive Control (VLM-PC), combining two key components that we find to be crucial for eliciting on-the-fly, adaptive behavior selection.
arXiv Detail & Related papers (2024-07-02T21:00:30Z) - Interactive Multi-Robot Flocking with Gesture Responsiveness and Musical Accompaniment [0.7659052547635159]
This work presents a compelling multi-robot task in which the main aim is to enthrall and interest.
In this task, the goal is for a human to be drawn to move alongside and participate in a dynamic, expressive robot flock.
Towards this aim, the research team created algorithms for robot movements and engaging interaction modes such as gestures and sound.
arXiv Detail & Related papers (2024-03-30T18:16:28Z) - HumanoidBench: Simulated Humanoid Benchmark for Whole-Body Locomotion and Manipulation [50.616995671367704]
We present a high-dimensional, simulated robot learning benchmark, HumanoidBench, featuring a humanoid robot equipped with dexterous hands.
Our findings reveal that state-of-the-art reinforcement learning algorithms struggle with most tasks, whereas a hierarchical learning approach achieves superior performance when supported by robust low-level policies.
arXiv Detail & Related papers (2024-03-15T17:45:44Z) - Self Model for Embodied Intelligence: Modeling Full-Body Human Musculoskeletal System and Locomotion Control with Hierarchical Low-Dimensional Representation [22.925312305575183]
We build a musculoskeletal model (MS-Human-700) with 90 body segments, 206 joints, and 700 muscle-tendon units.
We develop a new algorithm using low-dimensional representation and hierarchical deep reinforcement learning to achieve state-of-the-art full-body control.
arXiv Detail & Related papers (2023-12-09T05:42:32Z) - Persistent-Transient Duality: A Multi-mechanism Approach for Modeling
Human-Object Interaction [58.67761673662716]
Humans are highly adaptable, swiftly switching between different modes to handle different tasks, situations and contexts.
In Human-object interaction (HOI) activities, these modes can be attributed to two mechanisms: (1) the large-scale consistent plan for the whole activity and (2) the small-scale children interactive actions that start and end along the timeline.
This work proposes to model two concurrent mechanisms that jointly control human motion.
arXiv Detail & Related papers (2023-07-24T12:21:33Z) - Model Predictive Control for Fluid Human-to-Robot Handovers [50.72520769938633]
Planning motions that take human comfort into account is not a part of the human-robot handover process.
We propose to generate smooth motions via an efficient model-predictive control framework.
We conduct human-to-robot handover experiments on a diverse set of objects with several users.
arXiv Detail & Related papers (2022-03-31T23:08:20Z) - Synthesis and Execution of Communicative Robotic Movements with
Generative Adversarial Networks [59.098560311521034]
We focus on how to transfer on two different robotic platforms the same kinematics modulation that humans adopt when manipulating delicate objects.
We choose to modulate the velocity profile adopted by the robots' end-effector, inspired by what humans do when transporting objects with different characteristics.
We exploit a novel Generative Adversarial Network architecture, trained with human kinematics examples, to generalize over them and generate new and meaningful velocity profiles.
arXiv Detail & Related papers (2022-03-29T15:03:05Z) - Motron: Multimodal Probabilistic Human Motion Forecasting [30.154996245556532]
Motron is a graph-structured model that captures human's multimodality.
It outputs deterministic motions and corresponding confidence values for each mode.
We demonstrate the performance of our model on several challenging real-world motion forecasting datasets.
arXiv Detail & Related papers (2022-03-08T14:58:41Z) - Sensorimotor representation learning for an "active self" in robots: A
model survey [10.649413494649293]
In humans, these capabilities are thought to be related to our ability to perceive our body in space.
This paper reviews the developmental processes of underlying mechanisms of these abilities.
We propose a theoretical computational framework, which aims to allow the emergence of the sense of self in artificial agents.
arXiv Detail & Related papers (2020-11-25T16:31:01Z) - Learning Agile Robotic Locomotion Skills by Imitating Animals [72.36395376558984]
Reproducing the diverse and agile locomotion skills of animals has been a longstanding challenge in robotics.
We present an imitation learning system that enables legged robots to learn agile locomotion skills by imitating real-world animals.
arXiv Detail & Related papers (2020-04-02T02:56:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.