On the Critical Role of Conventions in Adaptive Human-AI Collaboration
- URL: http://arxiv.org/abs/2104.02871v1
- Date: Wed, 7 Apr 2021 02:46:19 GMT
- Title: On the Critical Role of Conventions in Adaptive Human-AI Collaboration
- Authors: Andy Shih and Arjun Sawhney and Jovana Kondic and Stefano Ermon and
Dorsa Sadigh
- Abstract summary: We propose a learning framework that teases apart rule-dependent representation from convention-dependent representation.
We experimentally validate our approach on three collaborative tasks varying in complexity.
- Score: 73.21967490610142
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Humans can quickly adapt to new partners in collaborative tasks (e.g. playing
basketball), because they understand which fundamental skills of the task (e.g.
how to dribble, how to shoot) carry over across new partners. Humans can also
quickly adapt to similar tasks with the same partners by carrying over
conventions that they have developed (e.g. raising hand signals pass the ball),
without learning to coordinate from scratch. To collaborate seamlessly with
humans, AI agents should adapt quickly to new partners and new tasks as well.
However, current approaches have not attempted to distinguish between the
complexities intrinsic to a task and the conventions used by a partner, and
more generally there has been little focus on leveraging conventions for
adapting to new settings. In this work, we propose a learning framework that
teases apart rule-dependent representation from convention-dependent
representation in a principled way. We show that, under some assumptions, our
rule-dependent representation is a sufficient statistic of the distribution
over best-response strategies across partners. Using this separation of
representations, our agents are able to adapt quickly to new partners, and to
coordinate with old partners on new tasks in a zero-shot manner. We
experimentally validate our approach on three collaborative tasks varying in
complexity: a contextual multi-armed bandit, a block placing task, and the card
game Hanabi.
Related papers
- ProAgent: Building Proactive Cooperative Agents with Large Language
Models [89.53040828210945]
ProAgent is a novel framework that harnesses large language models to create proactive agents.
ProAgent can analyze the present state, and infer the intentions of teammates from observations.
ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various coordination scenarios.
arXiv Detail & Related papers (2023-08-22T10:36:56Z) - Tackling Cooperative Incompatibility for Zero-Shot Human-AI Coordination [36.33334853998621]
We introduce the Cooperative Open-ended LEarning (COLE) framework to solve cooperative incompatibility in learning.
COLE formulates open-ended objectives in cooperative games with two players using perspectives of graph theory to evaluate and pinpoint the cooperative capacity of each strategy.
We show that COLE could effectively overcome the cooperative incompatibility from theoretical and empirical analysis.
arXiv Detail & Related papers (2023-06-05T16:51:38Z) - Adaptive Coordination in Social Embodied Rearrangement [49.35582108902819]
We study zero-shot coordination (ZSC) in this task, where an agent collaborates with a new partner, emulating a scenario where a robot collaborates with a new human partner.
We propose Behavior Diversity Play (BDP), a novel ZSC approach that encourages diversity through a discriminability objective.
Our results demonstrate that BDP learns adaptive agents that can tackle visual coordination, and zero-shot generalize to new partners in unseen environments, achieving 35% higher success and 32% higher efficiency compared to baselines.
arXiv Detail & Related papers (2023-05-31T18:05:51Z) - A Hierarchical Approach to Population Training for Human-AI
Collaboration [20.860808795671343]
We introduce a Hierarchical Reinforcement Learning (HRL) based method for Human-AI Collaboration.
We demonstrate that our method is able to dynamically adapt to novel partners of different play styles and skill levels in the 2-player collaborative Overcooked game environment.
arXiv Detail & Related papers (2023-05-26T07:53:12Z) - PECAN: Leveraging Policy Ensemble for Context-Aware Zero-Shot Human-AI
Coordination [52.991211077362586]
We propose a policy ensemble method to increase the diversity of partners in the population.
We then develop a context-aware method enabling the ego agent to analyze and identify the partner's potential policy primitives.
In this way, the ego agent is able to learn more universal cooperative behaviors for collaborating with diverse partners.
arXiv Detail & Related papers (2023-01-16T12:14:58Z) - Conditional Imitation Learning for Multi-Agent Games [89.897635970366]
We study the problem of conditional multi-agent imitation learning, where we have access to joint trajectory demonstrations at training time.
We propose a novel approach to address the difficulties of scalability and data scarcity.
Our model learns a low-rank subspace over ego and partner agent strategies, then infers and adapts to a new partner strategy by interpolating in the subspace.
arXiv Detail & Related papers (2022-01-05T04:40:13Z) - A Cordial Sync: Going Beyond Marginal Policies for Multi-Agent Embodied
Tasks [111.34055449929487]
We introduce the novel task FurnMove in which agents work together to move a piece of furniture through a living room to a goal.
Unlike existing tasks, FurnMove requires agents to coordinate at every timestep.
We identify two challenges when training agents to complete FurnMove: existing decentralized action sampling procedures do not permit expressive joint action policies.
Using SYNC-policies and CORDIAL, our agents achieve a 58% completion rate on FurnMove, an impressive absolute gain of 25 percentage points over competitive decentralized baselines.
arXiv Detail & Related papers (2020-07-09T17:59:57Z) - Evolving Dyadic Strategies for a Cooperative Physical Task [0.0]
We evolve simulated agents to explore a space of feasible role-switching policies.
Applying these switching policies in a cooperative manual task, agents process visual and haptic cues to decide when to switch roles.
We find that the best performing dyads exhibit high temporal coordination (anti-synchrony)
arXiv Detail & Related papers (2020-04-22T13:23:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.