Strategic Apple Tasting
- URL: http://arxiv.org/abs/2306.06250v2
- Date: Sat, 28 Oct 2023 03:21:47 GMT
- Title: Strategic Apple Tasting
- Authors: Keegan Harris, Chara Podimata, Zhiwei Steven Wu
- Abstract summary: Algorithmic decision-making in high-stakes domains often involves assigning decisions to agents with incentives to strategically modify their input to the algorithm.
We formalize this setting as an online learning problem with apple-tasting feedback.
Our goal is to achieve sublinear strategic regret, which compares the performance of the principal to that of the best fixed policy in hindsight.
- Score: 35.25249063553063
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Algorithmic decision-making in high-stakes domains often involves assigning
decisions to agents with incentives to strategically modify their input to the
algorithm. In addition to dealing with incentives, in many domains of interest
(e.g. lending and hiring) the decision-maker only observes feedback regarding
their policy for rounds in which they assign a positive decision to the agent;
this type of feedback is often referred to as apple tasting (or one-sided)
feedback. We formalize this setting as an online learning problem with
apple-tasting feedback where a principal makes decisions about a sequence of
$T$ agents, each of which is represented by a context that may be strategically
modified. Our goal is to achieve sublinear strategic regret, which compares the
performance of the principal to that of the best fixed policy in hindsight, if
the agents were truthful when revealing their contexts. Our main result is a
learning algorithm which incurs $O (\sqrt{T})$ strategic regret when the
sequence of agents is chosen stochastically. We also give an algorithm capable
of handling adversarially-chosen agents, albeit at the cost of
$O(T^{(d+1)/(d+2)})$ strategic regret (where $d$ is the dimension of the
context). Our algorithms can be easily adapted to the setting where the
principal receives bandit feedback -- this setting generalizes both the linear
contextual bandit problem (by considering agents with incentives) and the
strategic classification problem (by allowing for partial feedback).
Related papers
- Paths to Equilibrium in Games [6.812247730094933]
We study sequences of strategies satisfying a pairwise constraint inspired by policy updating in reinforcement learning.
Our analysis reveals a counterintuitive insight that reward deteriorating strategic updates are key to driving play to equilibrium along a satisficing path.
arXiv Detail & Related papers (2024-03-26T19:58:39Z) - Strategically-Robust Learning Algorithms for Bidding in First-Price Auctions [11.988955088595858]
Learning to bid in repeated first-price auctions is a fundamental problem at the interface of game theory and machine learning.
We propose a novel concave formulation for pure-strategy bidding in first-price auctions, and use it to analyze natural Gradient-Ascent-based algorithms for this problem.
arXiv Detail & Related papers (2024-02-12T01:33:33Z) - One-Shot Strategic Classification Under Unknown Costs [19.390528752448283]
We show that for a broad class of costs, even small mis-estimations of the cost function can entail trivial accuracy in the worst case.
Our analysis reveals important strategic responses, particularly the value of dual regularization with respect to the cost manipulation function.
arXiv Detail & Related papers (2023-11-05T20:43:08Z) - Learning to Incentivize Information Acquisition: Proper Scoring Rules
Meet Principal-Agent Model [64.94131130042275]
We study the incentivized information acquisition problem, where a principal hires an agent to gather information on her behalf.
We design a provably sample efficient algorithm that tailors the UCB algorithm to our model.
Our algorithm features a delicate estimation procedure for the optimal profit of the principal, and a conservative correction scheme that ensures the desired agent's actions are incentivized.
arXiv Detail & Related papers (2023-03-15T13:40:16Z) - Learning in Stackelberg Games with Non-myopic Agents [60.927889817803745]
We study Stackelberg games where a principal repeatedly interacts with a non-myopic long-lived agent, without knowing the agent's payoff function.
We provide a general framework that reduces learning in presence of non-myopic agents to robust bandit optimization in the presence of myopic agents.
arXiv Detail & Related papers (2022-08-19T15:49:30Z) - Decentralized Optimistic Hyperpolicy Mirror Descent: Provably No-Regret
Learning in Markov Games [95.10091348976779]
We study decentralized policy learning in Markov games where we control a single agent to play with nonstationary and possibly adversarial opponents.
We propose a new algorithm, underlineDecentralized underlineOptimistic hypeunderlineRpolicy munderlineIrror deunderlineScent (DORIS)
DORIS achieves $sqrtK$-regret in the context of general function approximation, where $K$ is the number of episodes.
arXiv Detail & Related papers (2022-06-03T14:18:05Z) - Who Leads and Who Follows in Strategic Classification? [82.44386576129295]
We argue that the order of play in strategic classification is fundamentally determined by the relative frequencies at which the decision-maker and the agents adapt to each other's actions.
We show that a decision-maker with the freedom to choose their update frequency can induce learning dynamics that converge to Stackelberg equilibria with either order of play.
arXiv Detail & Related papers (2021-06-23T16:48:46Z) - Optimally Deceiving a Learning Leader in Stackelberg Games [123.14187606686006]
Recent results in the ML community have revealed that learning algorithms used to compute the optimal strategy for the leader to commit to in a Stackelberg game, are susceptible to manipulation by the follower.
This paper shows that it is always possible for the follower to compute (near-optimal) payoffs for various scenarios about the learning interaction between leader and follower.
arXiv Detail & Related papers (2020-06-11T16:18:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.