Learning Recourse Costs from Pairwise Feature Comparisons
- URL: http://arxiv.org/abs/2409.13940v1
- Date: Fri, 20 Sep 2024 23:04:08 GMT
- Title: Learning Recourse Costs from Pairwise Feature Comparisons
- Authors: Kaivalya Rawal, Himabindu Lakkaraju,
- Abstract summary: This paper presents a novel technique for incorporating user input when learning and inferring user preferences.
We propose the use of the Bradley-Terry model to automatically infer feature-wise costs using non-exhaustive human comparison surveys.
We demonstrate the efficient learning of individual feature costs using MAP estimates, and show that these non-exhaustive human surveys are sufficient to learn an exhaustive set of feature costs.
- Score: 22.629956883958076
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents a novel technique for incorporating user input when learning and inferring user preferences. When trying to provide users of black-box machine learning models with actionable recourse, we often wish to incorporate their personal preferences about the ease of modifying each individual feature. These recourse finding algorithms usually require an exhaustive set of tuples associating each feature to its cost of modification. Since it is hard to obtain such costs by directly surveying humans, in this paper, we propose the use of the Bradley-Terry model to automatically infer feature-wise costs using non-exhaustive human comparison surveys. We propose that users only provide inputs comparing entire recourses, with all candidate feature modifications, determining which recourses are easier to implement relative to others, without explicit quantification of their costs. We demonstrate the efficient learning of individual feature costs using MAP estimates, and show that these non-exhaustive human surveys, which do not necessarily contain data for each feature pair comparison, are sufficient to learn an exhaustive set of feature costs, where each feature is associated with a modification cost.
Related papers
- Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback [87.37721254914476]
We introduce a routing framework that combines inputs from humans and LMs to achieve better annotation quality.
We train a performance prediction model to predict a reward model's performance on an arbitrary combination of human and LM annotations.
We show that the selected hybrid mixture achieves better reward model performance compared to using either one exclusively.
arXiv Detail & Related papers (2024-10-24T20:04:15Z) - Learn What You Need in Personalized Federated Learning [53.83081622573734]
$textitLearn2pFed$ is a novel algorithm-unrolling-based personalized federated learning framework.
We show that $textitLearn2pFed$ significantly outperforms previous personalized federated learning methods.
arXiv Detail & Related papers (2024-01-16T12:45:15Z) - Collaborative Anomaly Detection [66.51075412012581]
We propose collaborative anomaly detection (CAD) to jointly learn all tasks with an embedding encoding correlations among tasks.
We explore CAD with conditional density estimation and conditional likelihood ratio estimation.
It is beneficial to select a small number of tasks in advance to learn a task embedding model, and then use it to warm-start all task embeddings.
arXiv Detail & Related papers (2022-09-20T18:01:07Z) - One for All: Simultaneous Metric and Preference Learning over Multiple
Users [17.083305162005136]
We study the simultaneous preference and metric learning from a crowd of respondents.
Our model jointly learns a distance metric that characterizes the crowd's general measure of item similarities.
We demonstrate the performance of our model on both simulated data and on a dataset of color preference judgements.
arXiv Detail & Related papers (2022-07-07T22:47:13Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Personalized Algorithmic Recourse with Preference Elicitation [20.78332455864586]
We introduce PEAR, the first human-in-the-loop approach capable of providing personalized algorithmic recourse tailored to the needs of any end-user.
PEAR builds on insights from Bayesian Preference Elicitation to iteratively refine an estimate of the costs of actions by asking choice set queries to the target user.
Our empirical evaluation on real-world datasets highlights how PEAR produces high-quality personalized recourse in only a handful of iterations.
arXiv Detail & Related papers (2022-05-27T03:12:18Z) - Linear Speedup in Personalized Collaborative Learning [69.45124829480106]
Personalization in federated learning can improve the accuracy of a model for a user by trading off the model's bias.
We formalize the personalized collaborative learning problem as optimization of a user's objective.
We explore conditions under which we can optimally trade-off their bias for a reduction in variance.
arXiv Detail & Related papers (2021-11-10T22:12:52Z) - Low-Cost Algorithmic Recourse for Users With Uncertain Cost Functions [74.00030431081751]
We formalize the notion of user-specific cost functions and introduce a new method for identifying actionable recourses for users.
Our method satisfies up to 25.89 percentage points more users compared to strong baseline methods.
arXiv Detail & Related papers (2021-11-01T19:49:35Z) - Implications on Feature Detection when using the Benefit-Cost Ratio [0.0]
A popular trade-off choice is the ratio of both, the BCR (benefit-cost ratio)
In situations with large cost differences and small effect sizes, the BCR missed relevant features and preferred cheap noise features.
While the simple benefit-cost ratio offers an easy solution to incorporate costs, it is important to be aware of its risks.
arXiv Detail & Related papers (2020-08-12T08:25:42Z) - Active Preference Learning using Maximum Regret [10.317601896290467]
We study active preference learning as a framework for intuitively specifying the behaviour of autonomous robots.
In active preference learning, a user chooses the preferred behaviour from a set of alternatives, from which the robot learns the user's preferences.
arXiv Detail & Related papers (2020-05-08T14:31:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.