Listwise Learning to Rank with Deep Q-Networks
- URL: http://arxiv.org/abs/2002.07651v1
- Date: Thu, 13 Feb 2020 22:45:56 GMT
- Title: Listwise Learning to Rank with Deep Q-Networks
- Authors: Abhishek Sharma
- Abstract summary: We show that DeepQRank, our deep q-learning to rank agent, demonstrates performance that can be considered state-of-the-art.
We run our algorithm against Microsoft's LETOR listwise dataset and achieve an NDCG@1 of 0.5075, narrowly beating out the leading supervised learning model, SVMRank (0.4958)
- Score: 3.9726605190181976
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning to Rank is the problem involved with ranking a sequence of documents
based on their relevance to a given query. Deep Q-Learning has been shown to be
a useful method for training an agent in sequential decision making. In this
paper, we show that DeepQRank, our deep q-learning to rank agent, demonstrates
performance that can be considered state-of-the-art. Though less
computationally efficient than a supervised learning approach such as linear
regression, our agent has fewer limitations in terms of which format of data it
can use for training and evaluation. We run our algorithm against Microsoft's
LETOR listwise dataset and achieve an NDCG@1 (ranking accuracy in the range
[0,1]) of 0.5075, narrowly beating out the leading supervised learning model,
SVMRank (0.4958).
Related papers
- Zeroshot Listwise Learning to Rank Algorithm for Recommendation [5.694872363688119]
Learning to rank is a rare technology compared with other techniques such as deep neural networks.
We design a zeroshot listwise learning to rank algorithm for recommendation.
arXiv Detail & Related papers (2024-09-05T09:16:14Z) - Contextual Dual Learning Algorithm with Listwise Distillation for Unbiased Learning to Rank [26.69630281310365]
Unbiased Learning to Rank (ULTR) aims to leverage biased implicit user feedback (e.g., click) to optimize an unbiased ranking model.
We propose a Contextual Dual Learning Algorithm with Listwise Distillation (CDLA-LD) to address both position bias and contextual bias.
arXiv Detail & Related papers (2024-08-19T09:13:52Z) - Rank-LIME: Local Model-Agnostic Feature Attribution for Learning to Rank [16.780058676633914]
Rank-LIME is a model-agnostic, local, post-hoc linear feature attribution method for the task of learning to rank.
We employ novel correlation-based perturbations, differentiable ranking loss functions and introduce new metrics to evaluate ranking based additive feature attribution models.
arXiv Detail & Related papers (2022-12-24T12:14:32Z) - RankDNN: Learning to Rank for Few-shot Learning [70.49494297554537]
This paper introduces a new few-shot learning pipeline that casts relevance ranking for image retrieval as binary ranking relation classification.
It provides a new perspective on few-shot learning and is complementary to state-of-the-art methods.
arXiv Detail & Related papers (2022-11-28T13:59:31Z) - Learning Rate Curriculum [75.98230528486401]
We propose a novel curriculum learning approach termed Learning Rate Curriculum (LeRaC)
LeRaC uses a different learning rate for each layer of a neural network to create a data-agnostic curriculum during the initial training epochs.
We compare our approach with Curriculum by Smoothing (CBS), a state-of-the-art data-agnostic curriculum learning approach.
arXiv Detail & Related papers (2022-05-18T18:57:36Z) - GNNRank: Learning Global Rankings from Pairwise Comparisons via Directed
Graph Neural Networks [68.61934077627085]
We introduce GNNRank, a modeling framework compatible with any GNN capable of learning digraph embeddings.
We show that our methods attain competitive and often superior performance compared with existing approaches.
arXiv Detail & Related papers (2022-02-01T04:19:50Z) - MURAL: Meta-Learning Uncertainty-Aware Rewards for Outcome-Driven
Reinforcement Learning [65.52675802289775]
We show that an uncertainty aware classifier can solve challenging reinforcement learning problems.
We propose a novel method for computing the normalized maximum likelihood (NML) distribution.
We show that the resulting algorithm has a number of intriguing connections to both count-based exploration methods and prior algorithms for learning reward functions.
arXiv Detail & Related papers (2021-07-15T08:19:57Z) - PairRank: Online Pairwise Learning to Rank by Divide-and-Conquer [35.199462901346706]
We propose to estimate a pairwise learning to rank model online.
In each round, candidate documents are partitioned and ranked according to the model's confidence on the estimated pairwise rank order.
Regret directly defined on the number of mis-ordered pairs is proven, which connects the online solution's theoretical convergence with its expected ranking performance.
arXiv Detail & Related papers (2021-02-28T01:16:55Z) - PiRank: Learning To Rank via Differentiable Sorting [85.28916333414145]
We propose PiRank, a new class of differentiable surrogates for ranking.
We show that PiRank exactly recovers the desired metrics in the limit of zero temperature.
arXiv Detail & Related papers (2020-12-12T05:07:36Z) - Rethinking Few-Shot Image Classification: a Good Embedding Is All You
Need? [72.00712736992618]
We show that a simple baseline: learning a supervised or self-supervised representation on the meta-training set, outperforms state-of-the-art few-shot learning methods.
An additional boost can be achieved through the use of self-distillation.
We believe that our findings motivate a rethinking of few-shot image classification benchmarks and the associated role of meta-learning algorithms.
arXiv Detail & Related papers (2020-03-25T17:58:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.