MTAC: Hierarchical Reinforcement Learning-based Multi-gait
Terrain-adaptive Quadruped Controller
- URL: http://arxiv.org/abs/2401.03337v1
- Date: Wed, 1 Nov 2023 18:17:47 GMT
- Title: MTAC: Hierarchical Reinforcement Learning-based Multi-gait
Terrain-adaptive Quadruped Controller
- Authors: Nishaant Shah, Kshitij Tiwari, and Aniket Bera
- Abstract summary: Control of quadruped robots in dynamic and rough terrain environments is a challenging problem due to the many degrees of freedom of these robots.
Current locomotion controllers for quadrupeds are limited in their ability to produce multiple adaptive gaits, solve tasks in a time and resource-efficient manner, and require tedious training and manual tuning procedures.
We propose MTAC: a multi-gait terrain-adaptive controller, which utilizes a Hierarchical reinforcement learning (HRL) approach while being time and memory-efficient.
- Score: 12.300578189051963
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Urban search and rescue missions require rapid first response to minimize
loss of life and damage. Often, such efforts are assisted by humanitarian
robots which need to handle dynamic operational conditions such as uneven and
rough terrains, especially during mass casualty incidents like an earthquake.
Quadruped robots, owing to their versatile design, have the potential to assist
in such scenarios. However, control of quadruped robots in dynamic and rough
terrain environments is a challenging problem due to the many degrees of
freedom of these robots. Current locomotion controllers for quadrupeds are
limited in their ability to produce multiple adaptive gaits, solve tasks in a
time and resource-efficient manner, and require tedious training and manual
tuning procedures. To address these challenges, we propose MTAC: a multi-gait
terrain-adaptive controller, which utilizes a Hierarchical reinforcement
learning (HRL) approach while being time and memory-efficient. We show that our
proposed method scales well to a diverse range of environments with similar
compute times as state-of-the-art methods. Our method showed greater than 75%
on most tasks, outperforming previous work on the majority of test cases.
Related papers
- Commonsense Reasoning for Legged Robot Adaptation with Vision-Language Models [81.55156507635286]
Legged robots are physically capable of navigating a diverse variety of environments and overcoming a wide range of obstructions.
Current learning methods often struggle with generalization to the long tail of unexpected situations without heavy human supervision.
We propose a system, VLM-Predictive Control (VLM-PC), combining two key components that we find to be crucial for eliciting on-the-fly, adaptive behavior selection.
arXiv Detail & Related papers (2024-07-02T21:00:30Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - Learning and Adapting Agile Locomotion Skills by Transferring Experience [71.8926510772552]
We propose a framework for training complex robotic skills by transferring experience from existing controllers to jumpstart learning new tasks.
We show that our method enables learning complex agile jumping behaviors, navigating to goal locations while walking on hind legs, and adapting to new environments.
arXiv Detail & Related papers (2023-04-19T17:37:54Z) - Bi-Manual Manipulation and Attachment via Sim-to-Real Reinforcement
Learning [23.164743388342803]
We study how to solve bi-manual tasks using reinforcement learning trained in simulation.
We also discuss modifications to our simulated environment which lead to effective training of RL policies.
In this work, we design a Connect Task, where the aim is for two robot arms to pick up and attach two blocks with magnetic connection points.
arXiv Detail & Related papers (2022-03-15T21:49:20Z) - Autonomous Warehouse Robot using Deep Q-Learning [0.5138012450471438]
In warehouses, specialized agents need to navigate, avoid obstacles and maximize the use of space.
We propose using Deep Reinforcement Learning (DRL) to address the robot navigation and obstacle avoidance problem.
We use a strategic variation of Q-tables to perform multi-agent Q-learning.
arXiv Detail & Related papers (2022-02-21T07:16:51Z) - OSCAR: Data-Driven Operational Space Control for Adaptive and Robust
Robot Manipulation [50.59541802645156]
Operational Space Control (OSC) has been used as an effective task-space controller for manipulation.
We propose OSC for Adaptation and Robustness (OSCAR), a data-driven variant of OSC that compensates for modeling errors.
We evaluate our method on a variety of simulated manipulation problems, and find substantial improvements over an array of controller baselines.
arXiv Detail & Related papers (2021-10-02T01:21:38Z) - Lifelong Robotic Reinforcement Learning by Retaining Experiences [61.79346922421323]
Many multi-task reinforcement learning efforts assume the robot can collect data from all tasks at all times.
In this work, we study a practical sequential multi-task RL problem motivated by the practical constraints of physical robotic systems.
We derive an approach that effectively leverages the data and policies learned for previous tasks to cumulatively grow the robot's skill-set.
arXiv Detail & Related papers (2021-09-19T18:00:51Z) - Reinforcement Learning with Evolutionary Trajectory Generator: A General
Approach for Quadrupedal Locomotion [29.853927354893656]
We propose a novel RL-based approach that contains an evolutionary foot trajectory generator.
The generator continually optimize the shape of the output trajectory for the given task, providing diversified motion priors to guide the policy learning.
We deploy the controller learned in the simulation on a 12-DoF quadrupedal robot, and it can successfully traverse challenging scenarios with efficient gaits.
arXiv Detail & Related papers (2021-09-14T02:51:50Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z) - Proficiency Constrained Multi-Agent Reinforcement Learning for
Environment-Adaptive Multi UAV-UGV Teaming [2.745883395089022]
Mixed aerial and ground robot teams are widely used for disaster rescue, social security, precision agriculture, and military missions.
This paper developed a novel teaming method, proficiency aware multi-agent deep reinforcement learning (Mix-RL), to guide ground and aerial cooperation.
Mix-RL exploits robot capabilities while being aware of the adaptions of robot capabilities to task requirements and environment conditions.
arXiv Detail & Related papers (2020-02-10T16:19:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.