Multi-Principal Assistance Games: Definition and Collegial Mechanisms
- URL: http://arxiv.org/abs/2012.14536v1
- Date: Tue, 29 Dec 2020 00:06:47 GMT
- Title: Multi-Principal Assistance Games: Definition and Collegial Mechanisms
- Authors: Arnaud Fickinger, Simon Zhuang, Andrew Critch, Dylan Hadfield-Menell,
Stuart Russell
- Abstract summary: We introduce the concept of a multi-principal assistance game (MPAG)
In an MPAG, a single agent assists N human principals who may have widely different preferences.
We analyze in particular a generalization of apprenticeship learning in which the humans first perform some work to obtain utility and demonstrate their preferences.
- Score: 16.491889275389457
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce the concept of a multi-principal assistance game (MPAG), and
circumvent an obstacle in social choice theory, Gibbard's theorem, by using a
sufficiently collegial preference inference mechanism. In an MPAG, a single
agent assists N human principals who may have widely different preferences.
MPAGs generalize assistance games, also known as cooperative inverse
reinforcement learning games. We analyze in particular a generalization of
apprenticeship learning in which the humans first perform some work to obtain
utility and demonstrate their preferences, and then the robot acts to further
maximize the sum of human payoffs. We show in this setting that if the game is
sufficiently collegial, i.e. if the humans are responsible for obtaining a
sufficient fraction of the rewards through their own actions, then their
preferences are straightforwardly revealed through their work. This revelation
mechanism is non-dictatorial, does not limit the possible outcomes to two
alternatives, and is dominant-strategy incentive-compatible.
Related papers
- Learning to Assist Humans without Inferring Rewards [65.28156318196397]
We build upon prior work that studies assistance through the lens of empowerment.
An assistive agent aims to maximize the influence of the human's actions.
We prove that these representations estimate a similar notion of empowerment to that studied by prior work.
arXiv Detail & Related papers (2024-11-04T21:31:04Z) - Incentivized Learning in Principal-Agent Bandit Games [62.41639598376539]
This work considers a repeated principal-agent bandit game, where the principal can only interact with her environment through the agent.
The principal can influence the agent's decisions by offering incentives which add up to his rewards.
We present nearly optimal learning algorithms for the principal's regret in both multi-armed and linear contextual settings.
arXiv Detail & Related papers (2024-03-06T16:00:46Z) - Cooperation and Control in Delegation Games [1.3518297878940662]
We study multi-principal, multi-agent scenarios as delegation games.
In such games, there are two important failure modes: problems of control and problems of cooperation.
We show -- theoretically and empirically -- how these measures determine the principals' welfare.
arXiv Detail & Related papers (2024-02-24T14:17:41Z) - Optimal coordination in Minority Game: A solution from reinforcement
learning [6.0413802011767705]
The Minority Game is perhaps the simplest model that provides insights into how human coordinate to maximize the resource utilization.
Here, we turn to the paradigm of reinforcement learning, where individuals' strategies are evolving by evaluating both the past experience and rewards in the future.
We reveal that the population is able to reach the optimal allocation when individuals appreciate both the past experience and rewards in the future.
arXiv Detail & Related papers (2023-12-20T00:47:45Z) - Cooperation Dynamics in Multi-Agent Systems: Exploring Game-Theoretic Scenarios with Mean-Field Equilibria [0.0]
This paper investigates strategies to invoke cooperation in game-theoretic scenarios, namely the Iterated Prisoner's Dilemma.
Existing cooperative strategies are analyzed for their effectiveness in promoting group-oriented behavior in repeated games.
The study extends to scenarios with exponentially growing agent populations.
arXiv Detail & Related papers (2023-09-28T08:57:01Z) - Generating Personas for Games with Multimodal Adversarial Imitation
Learning [47.70823327747952]
Reinforcement learning has been widely successful in producing agents capable of playing games at a human level.
Going beyond reinforcement learning is necessary to model a wide range of human playstyles.
This paper presents a novel imitation learning approach to generate multiple persona policies for playtesting.
arXiv Detail & Related papers (2023-08-15T06:58:19Z) - End-to-End Learning and Intervention in Games [60.41921763076017]
We provide a unified framework for learning and intervention in games.
We propose two approaches, respectively based on explicit and implicit differentiation.
The analytical results are validated using several real-world problems.
arXiv Detail & Related papers (2020-10-26T18:39:32Z) - Multi-Principal Assistance Games [11.85513759444069]
Impossibility theorems in social choice theory and voting theory can be applied to such games.
We analyze in particular a bandit apprentice game in which the humans act first to demonstrate their individual preferences for the arms.
We propose a social choice method that uses shared control of a system to combine preference inference with social welfare optimization.
arXiv Detail & Related papers (2020-07-19T00:23:25Z) - Learning to Incentivize Other Learning Agents [73.03133692589532]
We show how to equip RL agents with the ability to give rewards directly to other agents, using a learned incentive function.
Such agents significantly outperform standard RL and opponent-shaping agents in challenging general-sum Markov games.
Our work points toward more opportunities and challenges along the path to ensure the common good in a multi-agent future.
arXiv Detail & Related papers (2020-06-10T20:12:38Z) - Intrinsic Motivation for Encouraging Synergistic Behavior [55.10275467562764]
We study the role of intrinsic motivation as an exploration bias for reinforcement learning in sparse-reward synergistic tasks.
Our key idea is that a good guiding principle for intrinsic motivation in synergistic tasks is to take actions which affect the world in ways that would not be achieved if the agents were acting on their own.
arXiv Detail & Related papers (2020-02-12T19:34:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.