Reducing Popularity Influence by Addressing Position Bias
- URL: http://arxiv.org/abs/2412.08780v1
- Date: Wed, 11 Dec 2024 21:16:37 GMT
- Title: Reducing Popularity Influence by Addressing Position Bias
- Authors: Andrii Dzhoha, Alexey Kurennoy, Vladimir Vlasov, Marjan Celikik,
- Abstract summary: We show that position debiasing can effectively reduce a skew in the popularity of items induced by the position bias through a feedback loop.
We show that position debiasing can significantly improve assortment utilization, without any degradation in user engagement or financial metrics.
- Score: 0.0
- License:
- Abstract: Position bias poses a persistent challenge in recommender systems, with much of the existing research focusing on refining ranking relevance and driving user engagement. However, in practical applications, the mitigation of position bias does not always result in detectable short-term improvements in ranking relevance. This paper provides an alternative, practically useful view of what position bias reduction methods can achieve. It demonstrates that position debiasing can spread visibility and interactions more evenly across the assortment, effectively reducing a skew in the popularity of items induced by the position bias through a feedback loop. We offer an explanation of how position bias affects item popularity. This includes an illustrative model of the item popularity histogram and the effect of the position bias on its skewness. Through offline and online experiments on our large-scale e-commerce platform, we show that position debiasing can significantly improve assortment utilization, without any degradation in user engagement or financial metrics. This makes the ranking fairer and helps attract more partners or content providers, benefiting the customers and the business in the long term.
Related papers
- ALBAR: Adversarial Learning approach to mitigate Biases in Action Recognition [52.537021302246664]
Action recognition models often suffer from background bias (i.e., inferring actions based on background cues) and foreground bias (i.e., relying on subject appearance)
We propose ALBAR, a novel adversarial training method that mitigates foreground and background biases without requiring specialized knowledge of the bias attributes.
We evaluate our method on established background and foreground bias protocols, setting a new state-of-the-art and strongly improving combined debiasing performance by over 12% on HMDB51.
arXiv Detail & Related papers (2025-01-31T20:47:06Z) - Popularity-Aware Alignment and Contrast for Mitigating Popularity Bias [34.006766098392525]
Collaborative Filtering (CF) typically suffers from the challenge of popularity bias due to the uneven distribution of items in real-world datasets.
This bias leads to a significant accuracy gap between popular and unpopular items.
We propose Popularity-Aware Alignment and Contrast (PAAC) to address two challenges.
arXiv Detail & Related papers (2024-05-31T09:14:48Z) - Going Beyond Popularity and Positivity Bias: Correcting for Multifactorial Bias in Recommender Systems [74.47680026838128]
Two typical forms of bias in user interaction data with recommender systems (RSs) are popularity bias and positivity bias.
We consider multifactorial selection bias affected by both item and rating value factors.
We propose smoothing and alternating gradient descent techniques to reduce variance and improve the robustness of its optimization.
arXiv Detail & Related papers (2024-04-29T12:18:21Z) - GPTBIAS: A Comprehensive Framework for Evaluating Bias in Large Language
Models [83.30078426829627]
Large language models (LLMs) have gained popularity and are being widely adopted by a large user community.
The existing evaluation methods have many constraints, and their results exhibit a limited degree of interpretability.
We propose a bias evaluation framework named GPTBIAS that leverages the high performance of LLMs to assess bias in models.
arXiv Detail & Related papers (2023-12-11T12:02:14Z) - Test Time Embedding Normalization for Popularity Bias Mitigation [6.145760252113906]
Popularity bias is a widespread problem in the field of recommender systems.
We propose 'Test Time Embedding Normalization' as a simple yet effective strategy for mitigating popularity bias.
arXiv Detail & Related papers (2023-08-22T08:57:44Z) - Ranking with Popularity Bias: User Welfare under Self-Amplification
Dynamics [19.59766711993837]
We propose and theoretically analyze a general mechanism by which item popularity, item quality, and position bias jointly impact user choice.
We show that naive popularity-biased recommenders induce linear regret by conflating item quality and popularity.
arXiv Detail & Related papers (2023-05-24T22:38:19Z) - Whole Page Unbiased Learning to Rank [59.52040055543542]
Unbiased Learning to Rank(ULTR) algorithms are proposed to learn an unbiased ranking model with biased click data.
We propose a Bias Agnostic whole-page unbiased Learning to rank algorithm, named BAL, to automatically find the user behavior model.
Experimental results on a real-world dataset verify the effectiveness of the BAL.
arXiv Detail & Related papers (2022-10-19T16:53:08Z) - Cross Pairwise Ranking for Unbiased Item Recommendation [57.71258289870123]
We develop a new learning paradigm named Cross Pairwise Ranking (CPR)
CPR achieves unbiased recommendation without knowing the exposure mechanism.
We prove in theory that this way offsets the influence of user/item propensity on the learning.
arXiv Detail & Related papers (2022-04-26T09:20:27Z) - Unbiased Pairwise Learning to Rank in Recommender Systems [4.058828240864671]
Unbiased learning to rank algorithms are appealing candidates and have already been applied in many applications with single categorical labels.
We propose a novel unbiased LTR algorithm to tackle the challenges, which innovatively models position bias in the pairwise fashion.
Experiment results on public benchmark datasets and internal live traffic show the superior results of the proposed method for both categorical and continuous labels.
arXiv Detail & Related papers (2021-11-25T06:04:59Z) - Incentives for Item Duplication under Fair Ranking Policies [69.14168955766847]
We study the behaviour of different fair ranking policies in the presence of duplicates.
We find that fairness-aware ranking policies may conflict with diversity, due to their potential to incentivize duplication more than policies solely focused on relevance.
arXiv Detail & Related papers (2021-10-29T11:11:15Z) - User-centered Evaluation of Popularity Bias in Recommender Systems [4.30484058393522]
Recommendation and ranking systems suffer from popularity bias; the tendency of the algorithm to favor a few popular items while under-representing the majority of other items.
In this paper, we show the limitations of the existing metrics to evaluate popularity bias mitigation when we want to assess these algorithms from the users' perspective.
We present an effective approach that mitigates popularity bias from the user-centered point of view.
arXiv Detail & Related papers (2021-03-10T22:12:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.