PairRank: Online Pairwise Learning to Rank by Divide-and-Conquer
- URL: http://arxiv.org/abs/2103.00368v2
- Date: Wed, 3 Mar 2021 05:42:50 GMT
- Title: PairRank: Online Pairwise Learning to Rank by Divide-and-Conquer
- Authors: Yiling Jia, Huazheng Wang, Stephen Guo, Hongning Wang
- Abstract summary: We propose to estimate a pairwise learning to rank model online.
In each round, candidate documents are partitioned and ranked according to the model's confidence on the estimated pairwise rank order.
Regret directly defined on the number of mis-ordered pairs is proven, which connects the online solution's theoretical convergence with its expected ranking performance.
- Score: 35.199462901346706
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Online Learning to Rank (OL2R) eliminates the need of explicit relevance
annotation by directly optimizing the rankers from their interactions with
users. However, the required exploration drives it away from successful
practices in offline learning to rank, which limits OL2R's empirical
performance and practical applicability. In this work, we propose to estimate a
pairwise learning to rank model online. In each round, candidate documents are
partitioned and ranked according to the model's confidence on the estimated
pairwise rank order, and exploration is only performed on the uncertain pairs
of documents, i.e., \emph{divide-and-conquer}. Regret directly defined on the
number of mis-ordered pairs is proven, which connects the online solution's
theoretical convergence with its expected ranking performance. Comparisons
against an extensive list of OL2R baselines on two public learning to rank
benchmark datasets demonstrate the effectiveness of the proposed solution.
Related papers
- TSPRank: Bridging Pairwise and Listwise Methods with a Bilinear Travelling Salesman Model [19.7255072094322]
Travelling Salesman Problem Rank (TSPRank) is a hybrid pairwise-listwise ranking method.
TSPRank's robustness and superior performance across different domains highlight its potential as a versatile and effective LETOR solution.
arXiv Detail & Related papers (2024-11-18T21:10:14Z) - Contextual Dual Learning Algorithm with Listwise Distillation for Unbiased Learning to Rank [26.69630281310365]
Unbiased Learning to Rank (ULTR) aims to leverage biased implicit user feedback (e.g., click) to optimize an unbiased ranking model.
We propose a Contextual Dual Learning Algorithm with Listwise Distillation (CDLA-LD) to address both position bias and contextual bias.
arXiv Detail & Related papers (2024-08-19T09:13:52Z) - Bipartite Ranking Fairness through a Model Agnostic Ordering Adjustment [54.179859639868646]
We propose a model agnostic post-processing framework xOrder for achieving fairness in bipartite ranking.
xOrder is compatible with various classification models and ranking fairness metrics, including supervised and unsupervised fairness metrics.
We evaluate our proposed algorithm on four benchmark data sets and two real-world patient electronic health record repositories.
arXiv Detail & Related papers (2023-07-27T07:42:44Z) - Unsupervised Dense Retrieval with Relevance-Aware Contrastive
Pre-Training [81.3781338418574]
We propose relevance-aware contrastive learning.
We consistently improve the SOTA unsupervised Contriever model on the BEIR and open-domain QA retrieval benchmarks.
Our method can not only beat BM25 after further pre-training on the target corpus but also serves as a good few-shot learner.
arXiv Detail & Related papers (2023-06-05T18:20:27Z) - Learning Neural Ranking Models Online from Implicit User Feedback [40.40829575021796]
We propose to learn a neural ranking model from users' implicit feedback (e.g., clicks) collected on the fly.
We focus on RankNet and LambdaRank, due to their great empirical success and wide adoption in offline settings.
arXiv Detail & Related papers (2022-01-17T23:11:39Z) - Calibrating Explore-Exploit Trade-off for Fair Online Learning to Rank [38.28889079095716]
Online learning to rank (OL2R) has attracted great research interests in recent years.
We propose a general framework to achieve fairness defined by group exposure in OL2R.
In particular, when the model is exploring a set of results for relevance feedback, we confine the exploration within a subset of random permutations.
arXiv Detail & Related papers (2021-11-01T07:22:05Z) - PiRank: Learning To Rank via Differentiable Sorting [85.28916333414145]
We propose PiRank, a new class of differentiable surrogates for ranking.
We show that PiRank exactly recovers the desired metrics in the limit of zero temperature.
arXiv Detail & Related papers (2020-12-12T05:07:36Z) - L2R2: Leveraging Ranking for Abductive Reasoning [65.40375542988416]
The abductive natural language inference task ($alpha$NLI) is proposed to evaluate the abductive reasoning ability of a learning system.
A novel $L2R2$ approach is proposed under the learning-to-rank framework.
Experiments on the ART dataset reach the state-of-the-art in the public leaderboard.
arXiv Detail & Related papers (2020-05-22T15:01:23Z) - Unbiased Learning to Rank: Online or Offline? [28.431648823968278]
How to obtain an unbiased ranking model by learning to rank with biased user feedback is an important research question for IR.
Existing work on unbiased learning to rank can be broadly categorized into two groups -- the studies on unbiased learning algorithms with logged data, and the studies on unbiased parameters estimation with real-time user interactions.
This paper formalizes the task of unbiased learning to rank and shows that existing algorithms for offline unbiased learning and online learning to rank are just the two sides of the same coin.
arXiv Detail & Related papers (2020-04-28T15:01:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.