A Non-Parametric Choice Model That Learns How Users Choose Between Recommended Options
- URL: http://arxiv.org/abs/2507.20035v1
- Date: Sat, 26 Jul 2025 18:38:27 GMT
- Title: A Non-Parametric Choice Model That Learns How Users Choose Between Recommended Options
- Authors: Thorsten Krause, Harrie Oosterhuis,
- Abstract summary: In recommendation settings, choice models consider which competitors appeared with the chosen item.<n>It is currently unclear how accurately these assumptions capture actual user behavior.<n>We show that learning choice models, instead of assuming them, can produce more robust predictions.
- Score: 15.763596979466929
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Choice models predict which items users choose from presented options. In recommendation settings, they can infer user preferences while countering exposure bias. In contrast with traditional univariate recommendation models, choice models consider which competitors appeared with the chosen item. This ability allows them to distinguish whether a user chose an item due to preference, i.e., they liked it; or competition, i.e., it was the best available option. Each choice model assumes specific user behavior, e.g., the multinomial logit model. However, it is currently unclear how accurately these assumptions capture actual user behavior, how wrong assumptions impact inference, and whether better models exist. In this work, we propose the learned choice model for recommendation (LCM4Rec), a non-parametric method for estimating the choice model. By applying kernel density estimation, LCM4Rec infers the most likely error distribution that describes the effect of inter-item cannibalization and thereby characterizes the users' choice model. Thus, it simultaneously infers what users prefer and how they make choices. Our experimental results indicate that our method (i) can accurately recover the choice model underlying a dataset; (ii) provides robust user preference inference, in contrast with existing choice models that are only effective when their assumptions match user behavior; and (iii) is more resistant against exposure bias than existing choice models. Thereby, we show that learning choice models, instead of assuming them, can produce more robust predictions. We believe this work provides an important step towards better understanding users' choice behavior.
Related papers
- Personalized Preference Fine-tuning of Diffusion Models [75.22218338096316]
We introduce PPD, a multi-reward optimization objective that aligns diffusion models with personalized preferences.<n>With PPD, a diffusion model learns the individual preferences of a population of users in a few-shot way.<n>Our approach achieves an average win rate of 76% over Stable Cascade, generating images that more accurately reflect specific user preferences.
arXiv Detail & Related papers (2025-01-11T22:38:41Z) - Beyond the Binary: Capturing Diverse Preferences With Reward Regularization [15.518838657050173]
We argue that this reliance on binary choices does not capture the broader, aggregate preferences of the target user in real-world tasks.<n>We introduce a simple yet effective method that augments existing binary preference datasets with synthetic preference judgments to estimate potential user disagreement.
arXiv Detail & Related papers (2024-12-05T02:35:46Z) - ComPO: Community Preferences for Language Model Personalization [122.54846260663922]
ComPO is a method to personalize preference optimization in language models.
We collect and release ComPRed, a question answering dataset with community-level preferences from Reddit.
arXiv Detail & Related papers (2024-10-21T14:02:40Z) - Large Language Models Are Not Robust Multiple Choice Selectors [117.72712117510953]
Multiple choice questions (MCQs) serve as a common yet important task format in the evaluation of large language models (LLMs)
This work shows that modern LLMs are vulnerable to option position changes due to their inherent "selection bias"
We propose a label-free, inference-time debiasing method, called PriDe, which separates the model's prior bias for option IDs from the overall prediction distribution.
arXiv Detail & Related papers (2023-09-07T17:44:56Z) - Probe: Learning Users' Personalized Projection Bias in Intertemporal
Choices [5.874142059884521]
In this work, we focus on two commonly observed biases: projection bias and the reference-point effect.
To address these biases, we propose a novel bias-embedded preference model called Probe.
The Probe incorporates a weight function to capture users' projection bias and a value function to account for the reference-point effect.
arXiv Detail & Related papers (2023-03-09T12:13:46Z) - Learning Choice Functions with Gaussian Processes [0.225596179391365]
In consumer theory, ranking available objects by means of preference relations yields the most common description of individual choices.
We propose a choice-model which allows an individual to express a set-valued choice.
arXiv Detail & Related papers (2023-02-01T12:46:43Z) - Latent User Intent Modeling for Sequential Recommenders [92.66888409973495]
Sequential recommender models learn to predict the next items a user is likely to interact with based on his/her interaction history on the platform.
Most sequential recommenders however lack a higher-level understanding of user intents, which often drive user behaviors online.
Intent modeling is thus critical for understanding users and optimizing long-term user experience.
arXiv Detail & Related papers (2022-11-17T19:00:24Z) - Rethinking Missing Data: Aleatoric Uncertainty-Aware Recommendation [59.500347564280204]
We propose a new Aleatoric Uncertainty-aware Recommendation (AUR) framework.
AUR consists of a new uncertainty estimator along with a normal recommender model.
As the chance of mislabeling reflects the potential of a pair, AUR makes recommendations according to the uncertainty.
arXiv Detail & Related papers (2022-09-22T04:32:51Z) - On the safe use of prior densities for Bayesian model selection [0.0]
We discuss the issue of prior sensitivity of the marginal likelihood and its role in model selection.
We also comment on the use of uninformative priors, which are very common choices in practice.
One of them involving a real-world application on exoplanet detection.
arXiv Detail & Related papers (2022-06-10T16:17:48Z) - On the estimation of discrete choice models to capture irrational
customer behaviors [4.683806391173103]
We show how to use partially-ranked preferences to efficiently model rational and irrational customer types from transaction data.
An extensive set of experiments assesses the predictive accuracy of the proposed approach.
arXiv Detail & Related papers (2021-09-08T19:19:51Z) - Probabilistic and Variational Recommendation Denoising [56.879165033014026]
Learning from implicit feedback is one of the most common cases in the application of recommender systems.
We propose probabilistic and variational recommendation denoising for implicit feedback.
We employ the proposed DPI and DVAE on four state-of-the-art recommendation models and conduct experiments on three datasets.
arXiv Detail & Related papers (2021-05-20T08:59:44Z) - Choice Set Confounding in Discrete Choice [29.25891648918572]
Existing learning methods overlook how choice set assignment affects the data.
We adapt methods from causal inference to the discrete choice setting.
We show that accounting for choice set confounding makes choices observed in hotel booking more consistent with rational utility-maximization.
arXiv Detail & Related papers (2021-05-17T15:39:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.