Coordination with Humans via Strategy Matching
- URL: http://arxiv.org/abs/2210.15099v1
- Date: Thu, 27 Oct 2022 01:00:50 GMT
- Title: Coordination with Humans via Strategy Matching
- Authors: Michelle Zhao, Reid Simmons, Henny Admoni
- Abstract summary: We present an algorithm for autonomously recognizing available task-completion strategies by observing human-human teams performing a collaborative task.
By transforming team actions into low dimensional representations using hidden Markov models, we can identify strategies without prior knowledge.
Robot policies are learned on each of the identified strategies to construct a Mixture-of-Experts model that adapts to the task strategies of unseen human partners.
- Score: 5.072077366588174
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Human and robot partners increasingly need to work together to perform tasks
as a team. Robots designed for such collaboration must reason about how their
task-completion strategies interplay with the behavior and skills of their
human team members as they coordinate on achieving joint goals. Our goal in
this work is to develop a computational framework for robot adaptation to human
partners in human-robot team collaborations. We first present an algorithm for
autonomously recognizing available task-completion strategies by observing
human-human teams performing a collaborative task. By transforming team actions
into low dimensional representations using hidden Markov models, we can
identify strategies without prior knowledge. Robot policies are learned on each
of the identified strategies to construct a Mixture-of-Experts model that
adapts to the task strategies of unseen human partners. We evaluate our model
on a collaborative cooking task using an Overcooked simulator. Results of an
online user study with 125 participants demonstrate that our framework improves
the task performance and collaborative fluency of human-agent teams, as
compared to state of the art reinforcement learning methods.
Related papers
- Decentralized and Lifelong-Adaptive Multi-Agent Collaborative Learning [57.652899266553035]
Decentralized and lifelong-adaptive multi-agent collaborative learning aims to enhance collaboration among multiple agents without a central server.
We propose DeLAMA, a decentralized multi-agent lifelong collaborative learning algorithm with dynamic collaboration graphs.
arXiv Detail & Related papers (2024-03-11T09:21:11Z) - Large Language Model-based Human-Agent Collaboration for Complex Task
Solving [94.3914058341565]
We introduce the problem of Large Language Models (LLMs)-based human-agent collaboration for complex task-solving.
We propose a Reinforcement Learning-based Human-Agent Collaboration method, ReHAC.
This approach includes a policy model designed to determine the most opportune stages for human intervention within the task-solving process.
arXiv Detail & Related papers (2024-02-20T11:03:36Z) - PECAN: Leveraging Policy Ensemble for Context-Aware Zero-Shot Human-AI
Coordination [52.991211077362586]
We propose a policy ensemble method to increase the diversity of partners in the population.
We then develop a context-aware method enabling the ego agent to analyze and identify the partner's potential policy primitives.
In this way, the ego agent is able to learn more universal cooperative behaviors for collaborating with diverse partners.
arXiv Detail & Related papers (2023-01-16T12:14:58Z) - Behaviour-conditioned policies for cooperative reinforcement learning
tasks [41.74498230885008]
In various real-world tasks, an agent needs to cooperate with unknown partner agent types.
Deep reinforcement learning models can be trained to deliver the required functionality but are known to suffer from sample inefficiency and slow learning.
We suggest a method, where we synthetically produce populations of agents with different behavioural patterns together with ground truth data of their behaviour.
We additionally suggest an agent architecture, which can efficiently use the generated data and gain the meta-learning capability.
arXiv Detail & Related papers (2021-10-04T09:16:41Z) - Co-GAIL: Learning Diverse Strategies for Human-Robot Collaboration [51.268988527778276]
We present a method for learning a human-robot collaboration policy from human-human collaboration demonstrations.
Our method co-optimizes a human policy and a robot policy in an interactive learning process.
arXiv Detail & Related papers (2021-08-13T03:14:43Z) - Show Me What You Can Do: Capability Calibration on Reachable Workspace
for Human-Robot Collaboration [83.4081612443128]
We show that a short calibration using REMP can effectively bridge the gap between what a non-expert user thinks a robot can reach and the ground-truth.
We show that this calibration procedure not only results in better user perception, but also promotes more efficient human-robot collaborations.
arXiv Detail & Related papers (2021-03-06T09:14:30Z) - Forming Human-Robot Cooperation for Tasks with General Goal using
Evolutionary Value Learning [9.053709318841232]
In Human-Robot Cooperation (HRC), the robot cooperates with humans to accomplish the task together.
Existing approaches assume the human has a specific goal during the cooperation, and the robot infers and acts toward it.
We present the Evolutionary Value Learning (EVL) approach to model the dynamics of the goal specification process in HRC.
arXiv Detail & Related papers (2020-12-19T20:27:09Z) - Joint Mind Modeling for Explanation Generation in Complex Human-Robot
Collaborative Tasks [83.37025218216888]
We propose a novel explainable AI (XAI) framework for achieving human-like communication in human-robot collaborations.
The robot builds a hierarchical mind model of the human user and generates explanations of its own mind as a form of communications.
Results show that the generated explanations of our approach significantly improves the collaboration performance and user perception of the robot.
arXiv Detail & Related papers (2020-07-24T23:35:03Z) - Human-Robot Team Coordination with Dynamic and Latent Human Task
Proficiencies: Scheduling with Learning Curves [0.0]
We introduce a novel resource coordination that enables robots to explore the relative strengths and learning abilities of their human teammates.
We generate and evaluate a robust schedule while discovering the latest individual worker proficiency.
Results indicate that scheduling strategies favoring exploration tend to be beneficial for human-robot collaboration.
arXiv Detail & Related papers (2020-07-03T19:44:22Z) - Learn Task First or Learn Human Partner First: A Hierarchical Task
Decomposition Method for Human-Robot Cooperation [11.387868752604986]
This work proposes a novel task decomposition method with a hierarchical reward mechanism that enables the robot to learn the hierarchical dynamic control task separately from learning the human partner's behavior.
The results show that the robot should learn the task first to achieve higher team performance and learn the human first to achieve higher learning efficiency.
arXiv Detail & Related papers (2020-03-01T04:41:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.