Opening the Black Box: Interpretable Remedies for Popularity Bias in Recommender Systems
- URL: http://arxiv.org/abs/2508.17297v1
- Date: Sun, 24 Aug 2025 10:59:56 GMT
- Title: Opening the Black Box: Interpretable Remedies for Popularity Bias in Recommender Systems
- Authors: Parviz Ahmadov, Masoud Mansoury,
- Abstract summary: Popularity bias is a well-known challenge in recommender systems, where a small number of popular items receive disproportionate attention.<n>This imbalance often results in reduced recommendation quality and unfair exposure of items.<n>We propose a post-hoc method using a Sparse Autoencoder to interpret and mitigate popularity bias in deep recommendation models.
- Score: 1.8692254863855962
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Popularity bias is a well-known challenge in recommender systems, where a small number of popular items receive disproportionate attention, while the majority of less popular items are largely overlooked. This imbalance often results in reduced recommendation quality and unfair exposure of items. Although existing mitigation techniques address this bias to some extent, they typically lack transparency in how they operate. In this paper, we propose a post-hoc method using a Sparse Autoencoder (SAE) to interpret and mitigate popularity bias in deep recommendation models. The SAE is trained to replicate a pre-trained model's behavior while enabling neuron-level interpretability. By introducing synthetic users with clear preferences for either popular or unpopular items, we identify neurons encoding popularity signals based on their activation patterns. We then adjust the activations of the most biased neurons to steer recommendations toward fairer exposure. Experiments on two public datasets using a sequential recommendation model show that our method significantly improves fairness with minimal impact on accuracy. Moreover, it offers interpretability and fine-grained control over the fairness-accuracy trade-off.
Related papers
- From Insight to Intervention: Interpretable Neuron Steering for Controlling Popularity Bias in Recommender Systems [1.8692254863855962]
Popularity bias is a pervasive challenge in recommender systems, where a few popular items dominate attention while the majority of less popular items remain underexposed.<n>In this paper, we propose a post-hoc approach, PopSteer, that leverages a Sparse Autoencoder to both interpret and mitigate popularity bias in recommendation models.<n> Experiments on three public datasets with a sequential recommendation model demonstrate that PopSteer significantly enhances fairness with minimal impact on accuracy, while providing interpretable insights and fine-grained control over the fairness-accuracy trade-off.
arXiv Detail & Related papers (2026-01-21T16:02:11Z) - The Unfairness of Multifactorial Bias in Recommendation [68.35079031029616]
Popularity bias and positivity bias are prominent sources of bias in recommender systems.<n>In this work, we examine how multifactorial bias influences item-side fairness.<n>We adapt a percentile-based rating transformation as a pre-processing strategy to mitigate multifactorial bias.
arXiv Detail & Related papers (2026-01-19T08:37:43Z) - PBiLoss: Popularity-Aware Regularization to Improve Fairness in Graph-Based Recommender Systems [1.0128808054306186]
We propose PBiLoss, a regularization-based loss function designed to counteract popularity bias in graph-based recommender models explicitly.<n>We show that PBiLoss significantly improves fairness, as demonstrated by reductions in the Popularity-Rank Correlation for Users (PRU) and Popularity-Rank Correlation for Items (PRI)
arXiv Detail & Related papers (2025-07-25T08:29:32Z) - Finding Interest Needle in Popularity Haystack: Improving Retrieval by Modeling Item Exposure [8.3095709445007]
We introduce an exposure-aware retrieval scoring approach, which explicitly models item exposure probability and adjusts retrieval-stage ranking at inference time.<n>We validate our approach through online A/B experiments in a real-world video recommendation system, demonstrating a 25% increase in uniquely retrieved items and a 40% reduction in the dominance of over-popular content.<n>Our results establish a scalable, deployable solution for mitigating popularity bias at the retrieval stage, offering a new paradigm for bias-aware personalization.
arXiv Detail & Related papers (2025-03-31T00:04:01Z) - Towards Popularity-Aware Recommendation: A Multi-Behavior Enhanced Framework with Orthogonality Constraint [4.137753517504481]
Top-$K$ recommendation involves inferring latent user preferences and generating personalized recommendations.<n>We present a textbfPopularity-aware top-$K$ recommendation algorithm integrating multi-behavior textbfSide textbfInformation.
arXiv Detail & Related papers (2024-12-26T11:06:49Z) - Correcting Popularity Bias in Recommender Systems via Item Loss Equalization [1.7771454131646311]
A small set of popular items dominate the recommendation results due to their high interaction rates.<n>This phenomenon disproportionately benefits users with mainstream tastes while neglecting those with niche interests.<n>We propose an in-processing approach to address this issue by intervening in the training process of recommendation models.
arXiv Detail & Related papers (2024-10-07T08:34:18Z) - Going Beyond Popularity and Positivity Bias: Correcting for Multifactorial Bias in Recommender Systems [74.47680026838128]
Two typical forms of bias in user interaction data with recommender systems (RSs) are popularity bias and positivity bias.
We consider multifactorial selection bias affected by both item and rating value factors.
We propose smoothing and alternating gradient descent techniques to reduce variance and improve the robustness of its optimization.
arXiv Detail & Related papers (2024-04-29T12:18:21Z) - Take Care of Your Prompt Bias! Investigating and Mitigating Prompt Bias in Factual Knowledge Extraction [56.17020601803071]
Recent research shows that pre-trained language models (PLMs) suffer from "prompt bias" in factual knowledge extraction.
This paper aims to improve the reliability of existing benchmarks by thoroughly investigating and mitigating prompt bias.
arXiv Detail & Related papers (2024-03-15T02:04:35Z) - Test Time Embedding Normalization for Popularity Bias Mitigation [6.145760252113906]
Popularity bias is a widespread problem in the field of recommender systems.
We propose 'Test Time Embedding Normalization' as a simple yet effective strategy for mitigating popularity bias.
arXiv Detail & Related papers (2023-08-22T08:57:44Z) - Self-supervised debiasing using low rank regularization [59.84695042540525]
Spurious correlations can cause strong biases in deep neural networks, impairing generalization ability.
We propose a self-supervised debiasing framework potentially compatible with unlabeled samples.
Remarkably, the proposed debiasing framework significantly improves the generalization performance of self-supervised learning baselines.
arXiv Detail & Related papers (2022-10-11T08:26:19Z) - Cross Pairwise Ranking for Unbiased Item Recommendation [57.71258289870123]
We develop a new learning paradigm named Cross Pairwise Ranking (CPR)
CPR achieves unbiased recommendation without knowing the exposure mechanism.
We prove in theory that this way offsets the influence of user/item propensity on the learning.
arXiv Detail & Related papers (2022-04-26T09:20:27Z) - Contrastive Learning for Debiased Candidate Generation in Large-Scale
Recommender Systems [84.3996727203154]
We show that a popular choice of contrastive loss is equivalent to reducing the exposure bias via inverse propensity weighting.
We further improve upon CLRec and propose Multi-CLRec, for accurate multi-intention aware bias reduction.
Our methods have been successfully deployed in Taobao, where at least four-month online A/B tests and offline analyses demonstrate its substantial improvements.
arXiv Detail & Related papers (2020-05-20T08:15:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.