How and Why to Manipulate Your Own Agent
- URL: http://arxiv.org/abs/2112.07640v1
- Date: Tue, 14 Dec 2021 18:35:32 GMT
- Title: How and Why to Manipulate Your Own Agent
- Authors: Yoav Kolumbus, Noam Nisan
- Abstract summary: We consider strategic settings where several users engage in a repeated online interaction, assisted by regret-minimizing agents that repeatedly play a "game" on their behalf.
We study the dynamics and average outcomes of the repeated game of the agents, and view it as inducing a meta-game between the users.
- Score: 5.634825161148484
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider strategic settings where several users engage in a repeated
online interaction, assisted by regret-minimizing agents that repeatedly play a
"game" on their behalf. We study the dynamics and average outcomes of the
repeated game of the agents, and view it as inducing a meta-game between the
users. Our main focus is on whether users can benefit in this meta-game from
"manipulating" their own agent by mis-reporting their parameters to it. We
formally define this "user-agent meta-game" model for general games, discuss
its properties under different notions of convergence of the dynamics of the
automated agents and analyze the equilibria induced on the users in 2x2 games
in which the dynamics converge to a single equilibrium.
Related papers
- Ranking Joint Policies in Dynamic Games using Evolutionary Dynamics [0.0]
It has been shown that the dynamics of agents' interactions, even in simple two-player games, are incapable of reaching Nash equilibria.
Our goal is to identify agents' joint strategies that result in stable behavior, being resistant to changes, while also accounting for agents' payoffs.
arXiv Detail & Related papers (2025-02-20T16:50:38Z) - On Corrigibility and Alignment in Multi Agent Games [0.0]
Corrigibility of autonomous agents is an under explored part of system design.
It has been suggested that uncertainty over the human preferences acts to keep the agents corrigible, even in the face of human irrationality.
We present a general framework for modelling corrigibility in a multi-agent setting as a 2 player game in which the agents always have a move in which they can ask the human for supervision.
arXiv Detail & Related papers (2025-01-09T16:44:38Z) - Who's Gaming the System? A Causally-Motivated Approach for Detecting Strategic Adaptation [12.528928000871405]
We consider a multi-agent setting where the goal is to identify the "worst offenders:" agents that are gaming most aggressively.
We introduce a framework in which each agent's tendency to game is parameterized via a scalar.
By recasting the problem as a causal effect estimation problem where different agents represent different "treatments," we prove that a ranking of all agents by their gaming parameters is identifiable.
arXiv Detail & Related papers (2024-12-02T22:07:48Z) - Toward Optimal LLM Alignments Using Two-Player Games [86.39338084862324]
In this paper, we investigate alignment through the lens of two-agent games, involving iterative interactions between an adversarial and a defensive agent.
We theoretically demonstrate that this iterative reinforcement learning optimization converges to a Nash Equilibrium for the game induced by the agents.
Experimental results in safety scenarios demonstrate that learning in such a competitive environment not only fully trains agents but also leads to policies with enhanced generalization capabilities for both adversarial and defensive agents.
arXiv Detail & Related papers (2024-06-16T15:24:50Z) - Incorporating Rivalry in Reinforcement Learning for a Competitive Game [65.2200847818153]
This work proposes a novel reinforcement learning mechanism based on the social impact of rivalry behavior.
Our proposed model aggregates objective and social perception mechanisms to derive a rivalry score that is used to modulate the learning of artificial agents.
arXiv Detail & Related papers (2022-08-22T14:06:06Z) - Reinforcement Learning Agents in Colonel Blotto [0.0]
We focus on a specific instance of agent-based models, which uses reinforcement learning (RL) to train the agent how to act in its environment.
We find that the RL agent handily beats a single opponent, and still performs quite well when the number of opponents are increased.
We also analyze the RL agent and look at what strategies it has arrived by looking at the actions that it has given the highest and lowest Q-values.
arXiv Detail & Related papers (2022-04-04T16:18:01Z) - Who Leads and Who Follows in Strategic Classification? [82.44386576129295]
We argue that the order of play in strategic classification is fundamentally determined by the relative frequencies at which the decision-maker and the agents adapt to each other's actions.
We show that a decision-maker with the freedom to choose their update frequency can induce learning dynamics that converge to Stackelberg equilibria with either order of play.
arXiv Detail & Related papers (2021-06-23T16:48:46Z) - Policy Fusion for Adaptive and Customizable Reinforcement Learning
Agents [137.86426963572214]
We show how to combine distinct behavioral policies to obtain a meaningful "fusion" policy.
We propose four different policy fusion methods for combining pre-trained policies.
We provide several practical examples and use-cases for how these methods are indeed useful for video game production and designers.
arXiv Detail & Related papers (2021-04-21T16:08:44Z) - Learning Latent Representations to Influence Multi-Agent Interaction [65.44092264843538]
We propose a reinforcement learning-based framework for learning latent representations of an agent's policy.
We show that our approach outperforms the alternatives and learns to influence the other agent.
arXiv Detail & Related papers (2020-11-12T19:04:26Z) - Equilibria for Games with Combined Qualitative and Quantitative
Objectives [15.590197778287616]
We study concurrent games in which each player is a process that is assumed to act independently and strategically.
Our main result is that deciding the existence of a strict epsilon Nash equilibrium in such games is 2ExpTime-complete.
arXiv Detail & Related papers (2020-08-13T01:56:24Z) - Moody Learners -- Explaining Competitive Behaviour of Reinforcement
Learning Agents [65.2200847818153]
In a competitive scenario, the agent does not only have a dynamic environment but also is directly affected by the opponents' actions.
Observing the Q-values of the agent is usually a way of explaining its behavior, however, do not show the temporal-relation between the selected actions.
arXiv Detail & Related papers (2020-07-30T11:30:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.