Towards Principled Learning for Re-ranking in Recommender Systems
- URL: http://arxiv.org/abs/2504.04188v1
- Date: Sat, 05 Apr 2025 14:14:36 GMT
- Title: Towards Principled Learning for Re-ranking in Recommender Systems
- Authors: Qunwei Li, Linghui Li, Jianbin Lin, Wenliang Zhong,
- Abstract summary: Two principles are proposed, including convergence consistency and adversarial consistency.<n>These two principles can be applied in the learning of a generic re-ranker and improve its performance.
- Score: 10.881339468480467
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As the final stage of recommender systems, re-ranking presents ordered item lists to users that best match their interests. It plays such a critical role and has become a trending research topic with much attention from both academia and industry. Recent advances of re-ranking are focused on attentive listwise modeling of interactions and mutual influences among items to be re-ranked. However, principles to guide the learning process of a re-ranker, and to measure the quality of the output of the re-ranker, have been always missing. In this paper, we study such principles to learn a good re-ranker. Two principles are proposed, including convergence consistency and adversarial consistency. These two principles can be applied in the learning of a generic re-ranker and improve its performance. We validate such a finding by various baseline methods over different datasets.
Related papers
- Revisiting Reciprocal Recommender Systems: Metrics, Formulation, and Method [60.364834418531366]
We propose five new evaluation metrics that comprehensively and accurately assess the performance of RRS.
We formulate the RRS from a causal perspective, formulating recommendations as bilateral interventions.
We introduce a reranking strategy to maximize matching outcomes, as measured by the proposed metrics.
arXiv Detail & Related papers (2024-08-19T07:21:02Z) - FIRST: Faster Improved Listwise Reranking with Single Token Decoding [56.727761901751194]
First, we introduce FIRST, a novel listwise LLM reranking approach leveraging the output logits of the first generated identifier to directly obtain a ranked ordering of the candidates.
Empirical results demonstrate that FIRST accelerates inference by 50% while maintaining a robust ranking performance with gains across the BEIR benchmark.
Our results show that LLM rerankers can provide a stronger distillation signal compared to cross-encoders, yielding substantial improvements in retriever recall after relevance feedback.
arXiv Detail & Related papers (2024-06-21T21:27:50Z) - Full Stage Learning to Rank: A Unified Framework for Multi-Stage Systems [40.199257203898846]
We propose an improved ranking principle for multi-stage systems, namely the Generalized Probability Ranking Principle (GPRP)
GPRP emphasizes both the selection bias in each stage of the system pipeline as well as the underlying interest of users.
Our core idea is to first estimate the selection bias in the subsequent stages and then learn a ranking model that best complies with the downstream modules' selection bias.
arXiv Detail & Related papers (2024-05-08T06:35:04Z) - Adapting Double Q-Learning for Continuous Reinforcement Learning [0.65268245109828]
We present a novel approach to the bias correction, similar in spirit to Double Q-Learning.
Our approach shows promising near-SOTA results on a small set of MuJoCo environments.
arXiv Detail & Related papers (2023-09-25T19:09:54Z) - Improving Recommendation Relevance by simulating User Interest [77.34726150561087]
We observe that recommendation "recency" can be straightforwardly and transparently maintained by iterative reduction of ranks of inactive items.
The basic idea behind this work is patented in a context of online recommendation systems.
arXiv Detail & Related papers (2023-02-03T03:35:28Z) - Incorporating Relevance Feedback for Information-Seeking Retrieval using
Few-Shot Document Re-Ranking [56.80065604034095]
We introduce a kNN approach that re-ranks documents based on their similarity with the query and the documents the user considers relevant.
To evaluate our different integration strategies, we transform four existing information retrieval datasets into the relevance feedback scenario.
arXiv Detail & Related papers (2022-10-19T16:19:37Z) - Learning to Rank For Push Notifications Using Pairwise Expected Regret [8.990318688557]
New paradigms of content consumption present new challenges for ranking methods.
We introduce a novel ranking loss based on weighting the pairwise loss between candidates by the expected regret incurred for misordering the pair.
We demonstrate that the proposed method can outperform prior methods both in a simulated environment and in a production experiment on a major social network.
arXiv Detail & Related papers (2022-01-19T16:12:25Z) - D2RLIR : an improved and diversified ranking function in interactive
recommendation systems based on deep reinforcement learning [0.3058685580689604]
This paper proposes a deep reinforcement learning based recommendation system by utilizing Actor-Critic architecture.
The proposed model is able to generate a diverse while relevance recommendation list based on the user's preferences.
arXiv Detail & Related papers (2021-10-28T13:11:29Z) - RocketQAv2: A Joint Training Method for Dense Passage Retrieval and
Passage Re-ranking [89.82301733609279]
We propose a novel joint training approach for dense passage retrieval and passage re-ranking.
A major contribution is that we introduce the dynamic listwise distillation, where we design a unified listwise training approach for both the retriever and the re-ranker.
During the dynamic distillation, the retriever and the re-ranker can be adaptively improved according to each other's relevance information.
arXiv Detail & Related papers (2021-10-14T13:52:55Z) - Self-Supervised Reinforcement Learning for Recommender Systems [77.38665506495553]
We propose self-supervised reinforcement learning for sequential recommendation tasks.
Our approach augments standard recommendation models with two output layers: one for self-supervised learning and the other for RL.
Based on such an approach, we propose two frameworks namely Self-Supervised Q-learning(SQN) and Self-Supervised Actor-Critic(SAC)
arXiv Detail & Related papers (2020-06-10T11:18:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.