Gaming Helps! Learning from Strategic Interactions in Natural Dynamics
- URL: http://arxiv.org/abs/2002.07024v3
- Date: Sun, 28 Feb 2021 15:45:05 GMT
- Title: Gaming Helps! Learning from Strategic Interactions in Natural Dynamics
- Authors: Yahav Bechavod, Katrina Ligett, Zhiwei Steven Wu, Juba Ziani
- Abstract summary: We consider an online regression setting in which individuals adapt to the regression model.
We find that such strategic manipulations may in fact help the learner recover the meaningful variables.
- Score: 31.670231315699237
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider an online regression setting in which individuals adapt to the
regression model: arriving individuals are aware of the current model, and
invest strategically in modifying their own features so as to improve the
predicted score that the current model assigns to them. Such feature
manipulation has been observed in various scenarios -- from credit assessment
to school admissions -- posing a challenge for the learner. Surprisingly, we
find that such strategic manipulations may in fact help the learner recover the
meaningful variables -- that is, the features that, when changed, affect the
true label (as opposed to non-meaningful features that have no effect). We show
that even simple behavior on the learner's part allows her to simultaneously i)
accurately recover the meaningful features, and ii) incentivize agents to
invest in these meaningful features, providing incentives for improvement.
Related papers
- Feature Responsiveness Scores: Model-Agnostic Explanations for Recourse [7.730963708373791]
Consumer protection rules mandate that we provide a list of "principal reasons" to consumers who receive adverse decisions.
In practice, lenders and employers identify principal reasons by returning the top-scoring features from a feature attribution method.
We show that standard attribution methods can mislead individuals by highlighting reasons without recourse.
We propose to address these issues by scoring features on the basis of responsiveness.
arXiv Detail & Related papers (2024-10-29T23:37:49Z) - Learning under Imitative Strategic Behavior with Unforeseeable Outcomes [14.80947863438795]
We propose a Stackelberg game to model the interplay between individuals and the decision-maker.
We show that the objective difference between the two can be decomposed into three interpretable terms.
arXiv Detail & Related papers (2024-05-03T00:53:58Z) - Classification Under Strategic Self-Selection [13.168262355330299]
We study the effects of self-selection on learning and the implications of learning on the composition of the self-selected population.
We propose a differentiable framework for learning under self-selective behavior, which can be optimized effectively.
arXiv Detail & Related papers (2024-02-23T11:37:56Z) - MERMAIDE: Learning to Align Learners using Model-Based Meta-Learning [62.065503126104126]
We study how a principal can efficiently and effectively intervene on the rewards of a previously unseen learning agent in order to induce desirable outcomes.
This is relevant to many real-world settings like auctions or taxation, where the principal may not know the learning behavior nor the rewards of real people.
We introduce MERMAIDE, a model-based meta-learning framework to train a principal that can quickly adapt to out-of-distribution agents.
arXiv Detail & Related papers (2023-04-10T15:44:50Z) - Causal Strategic Classification: A Tale of Two Shifts [11.929584800629675]
We show how strategic behavior and causal effects underlie two complementing forms of distribution shift.
We propose a learning algorithm that balances these two forces and over time, and permits end-to-end training.
arXiv Detail & Related papers (2023-02-13T11:35:59Z) - Feature Forgetting in Continual Representation Learning [48.89340526235304]
representations do not suffer from "catastrophic forgetting" even in plain continual learning, but little further fact is known about its characteristics.
We devise a protocol for evaluating representation in continual learning, and then use it to present an overview of the basic trends of continual representation learning.
To study the feature forgetting problem, we create a synthetic dataset to identify and visualize the prevalence of feature forgetting in neural networks.
arXiv Detail & Related papers (2022-05-26T13:38:56Z) - Generative Adversarial Reward Learning for Generalized Behavior Tendency
Inference [71.11416263370823]
We propose a generative inverse reinforcement learning for user behavioral preference modelling.
Our model can automatically learn the rewards from user's actions based on discriminative actor-critic network and Wasserstein GAN.
arXiv Detail & Related papers (2021-05-03T13:14:25Z) - PsiPhi-Learning: Reinforcement Learning with Demonstrations using
Successor Features and Inverse Temporal Difference Learning [102.36450942613091]
We propose an inverse reinforcement learning algorithm, called emphinverse temporal difference learning (ITD)
We show how to seamlessly integrate ITD with learning from online environment interactions, arriving at a novel algorithm for reinforcement learning with demonstrations, called $Psi Phi$-learning.
arXiv Detail & Related papers (2021-02-24T21:12:09Z) - Linear Classifiers that Encourage Constructive Adaptation [6.324366770332667]
We study the dynamics of prediction and adaptation as a two-stage game, and characterize optimal strategies for the model designer and its decision subjects.
In benchmarks on simulated and real-world datasets, we find that classifiers trained using our method maintain the accuracy of existing approaches while inducing higher levels of improvement and less manipulation.
arXiv Detail & Related papers (2020-10-31T20:35:32Z) - Ecological Reinforcement Learning [76.9893572776141]
We study the kinds of environment properties that can make learning under such conditions easier.
understanding how properties of the environment impact the performance of reinforcement learning agents can help us to structure our tasks in ways that make learning tractable.
arXiv Detail & Related papers (2020-06-22T17:55:03Z) - Value-driven Hindsight Modelling [68.658900923595]
Value estimation is a critical component of the reinforcement learning (RL) paradigm.
Model learning can make use of the rich transition structure present in sequences of observations, but this approach is usually not sensitive to the reward function.
We develop an approach for representation learning in RL that sits in between these two extremes.
This provides tractable prediction targets that are directly relevant for a task, and can thus accelerate learning the value function.
arXiv Detail & Related papers (2020-02-19T18:10:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.