One for All: Simultaneous Metric and Preference Learning over Multiple
Users
- URL: http://arxiv.org/abs/2207.03609v1
- Date: Thu, 7 Jul 2022 22:47:13 GMT
- Title: One for All: Simultaneous Metric and Preference Learning over Multiple
Users
- Authors: Gregory Canal, Blake Mason, Ramya Korlakai Vinayak, Robert Nowak
- Abstract summary: We study the simultaneous preference and metric learning from a crowd of respondents.
Our model jointly learns a distance metric that characterizes the crowd's general measure of item similarities.
We demonstrate the performance of our model on both simulated data and on a dataset of color preference judgements.
- Score: 17.083305162005136
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper investigates simultaneous preference and metric learning from a
crowd of respondents. A set of items represented by $d$-dimensional feature
vectors and paired comparisons of the form ``item $i$ is preferable to item
$j$'' made by each user is given. Our model jointly learns a distance metric
that characterizes the crowd's general measure of item similarities along with
a latent ideal point for each user reflecting their individual preferences.
This model has the flexibility to capture individual preferences, while
enjoying a metric learning sample cost that is amortized over the crowd. We
first study this problem in a noiseless, continuous response setting (i.e.,
responses equal to differences of item distances) to understand the fundamental
limits of learning. Next, we establish prediction error guarantees for noisy,
binary measurements such as may be collected from human respondents, and show
how the sample complexity improves when the underlying metric is low-rank.
Finally, we establish recovery guarantees under assumptions on the response
distribution. We demonstrate the performance of our model on both simulated
data and on a dataset of color preference judgements across a large number of
users.
Related papers
- Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback [87.37721254914476]
We introduce a routing framework that combines inputs from humans and LMs to achieve better annotation quality.
We train a performance prediction model to predict a reward model's performance on an arbitrary combination of human and LM annotations.
We show that the selected hybrid mixture achieves better reward model performance compared to using either one exclusively.
arXiv Detail & Related papers (2024-10-24T20:04:15Z) - Learning Recourse Costs from Pairwise Feature Comparisons [22.629956883958076]
This paper presents a novel technique for incorporating user input when learning and inferring user preferences.
We propose the use of the Bradley-Terry model to automatically infer feature-wise costs using non-exhaustive human comparison surveys.
We demonstrate the efficient learning of individual feature costs using MAP estimates, and show that these non-exhaustive human surveys are sufficient to learn an exhaustive set of feature costs.
arXiv Detail & Related papers (2024-09-20T23:04:08Z) - Revisiting the Evaluation of Image Synthesis with GANs [55.72247435112475]
This study presents an empirical investigation into the evaluation of synthesis performance, with generative adversarial networks (GANs) as a representative of generative models.
In particular, we make in-depth analyses of various factors, including how to represent a data point in the representation space, how to calculate a fair distance using selected samples, and how many instances to use from each set.
arXiv Detail & Related papers (2023-04-04T17:54:32Z) - Non-Stochastic CDF Estimation Using Threshold Queries [3.6576781735746513]
We tackle the problem of estimating an empirical distribution in a setting with two challenges.
First, the algorithm does not directly observe the data; instead, it only asks a limited number of threshold queries about each sample.
Second, the data are not assumed to be independent and identically distributed; instead, we allow for an arbitrary process generating the samples.
arXiv Detail & Related papers (2023-01-13T18:00:57Z) - SeedBERT: Recovering Annotator Rating Distributions from an Aggregated
Label [43.23903984174963]
We propose SeedBERT, a method for recovering annotator rating distributions from a single label.
Our human evaluations indicate that SeedBERT's attention mechanism is consistent with human sources of annotator disagreement.
arXiv Detail & Related papers (2022-11-23T18:35:15Z) - Utilizing supervised models to infer consensus labels and their quality
from data with multiple annotators [16.79939549201032]
Real-world data for classification is often labeled by multiple annotators.
We introduce CROWDLAB, a straightforward approach to estimate such data.
Our proposed method provides superior estimates for (1)- (3) than many alternative algorithms.
arXiv Detail & Related papers (2022-10-13T07:54:07Z) - The Minority Matters: A Diversity-Promoting Collaborative Metric
Learning Algorithm [154.47590401735323]
Collaborative Metric Learning (CML) has recently emerged as a popular method in recommendation systems.
This paper focuses on a challenging scenario where a user has multiple categories of interests.
We propose a novel method called textitDiversity-Promoting Collaborative Metric Learning (DPCML)
arXiv Detail & Related papers (2022-09-30T08:02:18Z) - Learning Personalized Item-to-Item Recommendation Metric via Implicit
Feedback [24.37151414523712]
This paper studies the item-to-item recommendation problem in recommender systems from a new perspective of metric learning via implicit feedback.
We develop and investigate a personalizable deep metric model that captures both the internal contents of items and how they were interacted with by users.
arXiv Detail & Related papers (2022-03-18T18:08:57Z) - A Case Study on Sampling Strategies for Evaluating Neural Sequential
Item Recommendation Models [69.32128532935403]
Two well-known strategies to sample negative items are uniform random sampling and sampling by popularity.
We re-evaluate current state-of-the-art sequential recommender models from the point of view.
We find that both sampling strategies can produce inconsistent rankings compared with the full ranking of the models.
arXiv Detail & Related papers (2021-07-27T19:06:03Z) - Set2setRank: Collaborative Set to Set Ranking for Implicit Feedback
based Recommendation [59.183016033308014]
In this paper, we explore the unique characteristics of the implicit feedback and propose Set2setRank framework for recommendation.
Our proposed framework is model-agnostic and can be easily applied to most recommendation prediction approaches.
arXiv Detail & Related papers (2021-05-16T08:06:22Z) - One for More: Selecting Generalizable Samples for Generalizable ReID
Model [92.40951770273972]
This paper proposes a one-for-more training objective that takes the generalization ability of selected samples as a loss function.
Our proposed one-for-more based sampler can be seamlessly integrated into the ReID training framework.
arXiv Detail & Related papers (2020-12-10T06:37:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.