Optimizing Risk-averse Human-AI Hybrid Teams
- URL: http://arxiv.org/abs/2403.08386v1
- Date: Wed, 13 Mar 2024 09:49:26 GMT
- Title: Optimizing Risk-averse Human-AI Hybrid Teams
- Authors: Andrew Fuchs, Andrea Passarella, and Marco Conti
- Abstract summary: We propose a manager which learns, through a standard Reinforcement Learning scheme, how to best delegate.
We demonstrate the optimality of our manager's performance in several grid environments.
Our results show our manager can successfully learn desirable delegations which result in team paths near/exactly optimal.
- Score: 1.433758865948252
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We anticipate increased instances of humans and AI systems working together
in what we refer to as a hybrid team. The increase in collaboration is expected
as AI systems gain proficiency and their adoption becomes more widespread.
However, their behavior is not error-free, making hybrid teams a very suitable
solution. As such, we consider methods for improving performance for these
teams of humans and AI systems. For hybrid teams, we will refer to both the
humans and AI systems as agents. To improve team performance over that seen for
agents operating individually, we propose a manager which learns, through a
standard Reinforcement Learning scheme, how to best delegate, over time, the
responsibility of taking a decision to any of the agents. We further guide the
manager's learning so they also minimize how many changes in delegation are
made resulting from undesirable team behavior. We demonstrate the optimality of
our manager's performance in several grid environments which include failure
states which terminate an episode and should be avoided. We perform our
experiments with teams of agents with varying degrees of acceptable risk, in
the form of proximity to a failure state, and measure the manager's ability to
make effective delegation decisions with respect to its own risk-based
constraints, then compare these to the optimal decisions. Our results show our
manager can successfully learn desirable delegations which result in team paths
near/exactly optimal with respect to path length and number of delegations.
Related papers
- Symmetry-Breaking Augmentations for Ad Hoc Teamwork [10.014956508924842]
In many collaborative settings, artificial intelligence (AI) agents must be able to adapt to new teammates that use unknown or previously unobserved strategies.
We introduce symmetry-breaking augmentations (SBA), which increases diversity in the behaviour of training teammates by applying a symmetry-flipping operation.
We demonstrate this experimentally in two settings, and show that our approach improves upon previous ad hoc teamwork results in the challenging card game Hanabi.
arXiv Detail & Related papers (2024-02-15T14:49:28Z) - Optimizing Delegation in Collaborative Human-AI Hybrid Teams [1.433758865948252]
We address the case of hybrid teams in which only one team member (the control agent) is authorized to act as control for the team.
We propose the addition of an AI manager which learns as an outside observer of the team.
We demonstrate our manager performance in a simulated driving scenario representing the case of a hybrid team of agents composed of a human driver and autonomous driving system.
arXiv Detail & Related papers (2024-02-08T12:04:43Z) - Optimizing delegation between human and AI collaborative agents [1.6114012813668932]
We train a delegating manager agent to make delegation decisions with respect to potential performance deficiencies.
Our framework learns through observations of team performance without restricting agents to matching dynamics.
Our results show our manager learns to perform delegation decisions with teams of agents operating under differing representations of the environment.
arXiv Detail & Related papers (2023-09-26T07:23:26Z) - ProAgent: Building Proactive Cooperative Agents with Large Language
Models [89.53040828210945]
ProAgent is a novel framework that harnesses large language models to create proactive agents.
ProAgent can analyze the present state, and infer the intentions of teammates from observations.
ProAgent exhibits a high degree of modularity and interpretability, making it easily integrated into various coordination scenarios.
arXiv Detail & Related papers (2023-08-22T10:36:56Z) - Compensating for Sensing Failures via Delegation in Human-AI Hybrid
Systems [0.0]
We consider the hybrid human-AI teaming case where a managing agent is tasked with identifying when to perform a delegation assignment.
We model how the environmental context can contribute to, or exacerbate, the sensing deficiencies.
We demonstrate how a Reinforcement Learning (RL) manager can correct the context-delegation association.
arXiv Detail & Related papers (2023-03-02T14:27:01Z) - Coach-assisted Multi-Agent Reinforcement Learning Framework for
Unexpected Crashed Agents [120.91291581594773]
We present a formal formulation of a cooperative multi-agent reinforcement learning system with unexpected crashes.
We propose a coach-assisted multi-agent reinforcement learning framework, which introduces a virtual coach agent to adjust the crash rate during training.
To the best of our knowledge, this work is the first to study the unexpected crashes in the multi-agent system.
arXiv Detail & Related papers (2022-03-16T08:22:45Z) - Explore and Control with Adversarial Surprise [78.41972292110967]
Reinforcement learning (RL) provides a framework for learning goal-directed policies given user-specified rewards.
We propose a new unsupervised RL technique based on an adversarial game which pits two policies against each other to compete over the amount of surprise an RL agent experiences.
We show that our method leads to the emergence of complex skills by exhibiting clear phase transitions.
arXiv Detail & Related papers (2021-07-12T17:58:40Z) - Coach-Player Multi-Agent Reinforcement Learning for Dynamic Team
Composition [88.26752130107259]
In real-world multiagent systems, agents with different capabilities may join or leave without altering the team's overarching goals.
We propose COPA, a coach-player framework to tackle this problem.
We 1) adopt the attention mechanism for both the coach and the players; 2) propose a variational objective to regularize learning; and 3) design an adaptive communication method to let the coach decide when to communicate with the players.
arXiv Detail & Related papers (2021-05-18T17:27:37Z) - Learning to Incentivize Other Learning Agents [73.03133692589532]
We show how to equip RL agents with the ability to give rewards directly to other agents, using a learned incentive function.
Such agents significantly outperform standard RL and opponent-shaping agents in challenging general-sum Markov games.
Our work points toward more opportunities and challenges along the path to ensure the common good in a multi-agent future.
arXiv Detail & Related papers (2020-06-10T20:12:38Z) - Is the Most Accurate AI the Best Teammate? Optimizing AI for Teamwork [54.309495231017344]
We argue that AI systems should be trained in a human-centered manner, directly optimized for team performance.
We study this proposal for a specific type of human-AI teaming, where the human overseer chooses to either accept the AI recommendation or solve the task themselves.
Our experiments with linear and non-linear models on real-world, high-stakes datasets show that the most accuracy AI may not lead to highest team performance.
arXiv Detail & Related papers (2020-04-27T19:06:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.