Learn Task First or Learn Human Partner First: A Hierarchical Task
Decomposition Method for Human-Robot Cooperation
- URL: http://arxiv.org/abs/2003.00400v3
- Date: Tue, 7 Dec 2021 17:19:57 GMT
- Title: Learn Task First or Learn Human Partner First: A Hierarchical Task
Decomposition Method for Human-Robot Cooperation
- Authors: Lingfeng Tao, Michael Bowman, Jiucai Zhang, Xiaoli Zhang
- Abstract summary: This work proposes a novel task decomposition method with a hierarchical reward mechanism that enables the robot to learn the hierarchical dynamic control task separately from learning the human partner's behavior.
The results show that the robot should learn the task first to achieve higher team performance and learn the human first to achieve higher learning efficiency.
- Score: 11.387868752604986
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Applying Deep Reinforcement Learning (DRL) to Human-Robot Cooperation (HRC)
in dynamic control problems is promising yet challenging as the robot needs to
learn the dynamics of the controlled system and dynamics of the human partner.
In existing research, the robot powered by DRL adopts coupled observation of
the environment and the human partner to learn both dynamics simultaneously.
However, such a learning strategy is limited in terms of learning efficiency
and team performance. This work proposes a novel task decomposition method with
a hierarchical reward mechanism that enables the robot to learn the
hierarchical dynamic control task separately from learning the human partner's
behavior. The method is validated with a hierarchical control task in a
simulated environment with human subject experiments. Our method also provides
insight into the design of the learning strategy for HRC. The results show that
the robot should learn the task first to achieve higher team performance and
learn the human first to achieve higher learning efficiency.
Related papers
- SPIRE: Synergistic Planning, Imitation, and Reinforcement Learning for Long-Horizon Manipulation [58.14969377419633]
We propose spire, a system that decomposes tasks into smaller learning subproblems and second combines imitation and reinforcement learning to maximize their strengths.
We find that spire outperforms prior approaches that integrate imitation learning, reinforcement learning, and planning by 35% to 50% in average task performance.
arXiv Detail & Related papers (2024-10-23T17:42:07Z) - Large Language Model-based Human-Agent Collaboration for Complex Task
Solving [94.3914058341565]
We introduce the problem of Large Language Models (LLMs)-based human-agent collaboration for complex task-solving.
We propose a Reinforcement Learning-based Human-Agent Collaboration method, ReHAC.
This approach includes a policy model designed to determine the most opportune stages for human intervention within the task-solving process.
arXiv Detail & Related papers (2024-02-20T11:03:36Z) - Stabilizing Contrastive RL: Techniques for Robotic Goal Reaching from
Offline Data [101.43350024175157]
Self-supervised learning has the potential to decrease the amount of human annotation and engineering effort required to learn control strategies.
Our work builds on prior work showing that the reinforcement learning (RL) itself can be cast as a self-supervised problem.
We demonstrate that a self-supervised RL algorithm based on contrastive learning can solve real-world, image-based robotic manipulation tasks.
arXiv Detail & Related papers (2023-06-06T01:36:56Z) - Learning Human-to-Robot Handovers from Point Clouds [63.18127198174958]
We propose the first framework to learn control policies for vision-based human-to-robot handovers.
We show significant performance gains over baselines on a simulation benchmark, sim-to-sim transfer and sim-to-real transfer.
arXiv Detail & Related papers (2023-03-30T17:58:36Z) - Coordination with Humans via Strategy Matching [5.072077366588174]
We present an algorithm for autonomously recognizing available task-completion strategies by observing human-human teams performing a collaborative task.
By transforming team actions into low dimensional representations using hidden Markov models, we can identify strategies without prior knowledge.
Robot policies are learned on each of the identified strategies to construct a Mixture-of-Experts model that adapts to the task strategies of unseen human partners.
arXiv Detail & Related papers (2022-10-27T01:00:50Z) - Increased Complexity of a Human-Robot Collaborative Task May Increase
the Need for a Socially Competent Robot [0.0]
This study investigates how task complexity affects human perception and acceptance of their robot partner.
We propose a human-based robot control model for obstacle avoidance that can account for the leader-follower dynamics.
arXiv Detail & Related papers (2022-07-11T11:43:27Z) - Co-GAIL: Learning Diverse Strategies for Human-Robot Collaboration [51.268988527778276]
We present a method for learning a human-robot collaboration policy from human-human collaboration demonstrations.
Our method co-optimizes a human policy and a robot policy in an interactive learning process.
arXiv Detail & Related papers (2021-08-13T03:14:43Z) - Show Me What You Can Do: Capability Calibration on Reachable Workspace
for Human-Robot Collaboration [83.4081612443128]
We show that a short calibration using REMP can effectively bridge the gap between what a non-expert user thinks a robot can reach and the ground-truth.
We show that this calibration procedure not only results in better user perception, but also promotes more efficient human-robot collaborations.
arXiv Detail & Related papers (2021-03-06T09:14:30Z) - Forming Human-Robot Cooperation for Tasks with General Goal using
Evolutionary Value Learning [9.053709318841232]
In Human-Robot Cooperation (HRC), the robot cooperates with humans to accomplish the task together.
Existing approaches assume the human has a specific goal during the cooperation, and the robot infers and acts toward it.
We present the Evolutionary Value Learning (EVL) approach to model the dynamics of the goal specification process in HRC.
arXiv Detail & Related papers (2020-12-19T20:27:09Z) - Active Hierarchical Imitation and Reinforcement Learning [0.0]
In this project, we explored different imitation learning algorithms and designed active learning algorithms upon the hierarchical imitation and reinforcement learning framework we have developed.
Our experimental results showed that using DAgger and reward-based active learning method can achieve better performance while saving more human efforts physically and mentally during the training process.
arXiv Detail & Related papers (2020-12-14T08:27:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.