Listwise Learning to Rank by Exploring Unique Ratings
- URL: http://arxiv.org/abs/2001.01828v3
- Date: Thu, 23 Jan 2020 01:55:15 GMT
- Title: Listwise Learning to Rank by Exploring Unique Ratings
- Authors: Xiaofeng Zhu, Diego Klabjan
- Abstract summary: Existing listwise learning-to-rank models are generally derived from the classical Plackett-Luce model, which has three major limitations.
We propose a novel and efficient way of refining prediction scores by combining an adapted Vanilla Recurrent Neural Network (RNN) model with pooling given documents at previous steps.
Experiments demonstrate that the models notably outperform state-of-the-art learning-to-rank models.
- Score: 32.857847595096025
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose new listwise learning-to-rank models that mitigate
the shortcomings of existing ones. Existing listwise learning-to-rank models
are generally derived from the classical Plackett-Luce model, which has three
major limitations. (1) Its permutation probabilities overlook ties, i.e., a
situation when more than one document has the same rating with respect to a
query. This can lead to imprecise permutation probabilities and inefficient
training because of selecting documents one by one. (2) It does not favor
documents having high relevance. (3) It has a loose assumption that sampling
documents at different steps is independent. To overcome the first two
limitations, we model ranking as selecting documents from a candidate set based
on unique rating levels in decreasing order. The number of steps in training is
determined by the number of unique rating levels. We propose a new loss
function and associated four models for the entire sequence of weighted
classification tasks by assigning high weights to the selected documents with
high ratings for optimizing Normalized Discounted Cumulative Gain (NDCG). To
overcome the final limitation, we further propose a novel and efficient way of
refining prediction scores by combining an adapted Vanilla Recurrent Neural
Network (RNN) model with pooling given selected documents at previous steps. We
encode all of the documents already selected by an RNN model. In a single step,
we rank all of the documents with the same ratings using the last cell of the
RNN multiple times. We have implemented our models using three settings: neural
networks, neural networks with gradient boosting, and regression trees with
gradient boosting. We have conducted experiments on four public datasets. The
experiments demonstrate that the models notably outperform state-of-the-art
learning-to-rank models.
Related papers
- List-aware Reranking-Truncation Joint Model for Search and
Retrieval-augmented Generation [80.12531449946655]
We propose a Reranking-Truncation joint model (GenRT) that can perform the two tasks concurrently.
GenRT integrates reranking and truncation via generative paradigm based on encoder-decoder architecture.
Our method achieves SOTA performance on both reranking and truncation tasks for web search and retrieval-augmented LLMs.
arXiv Detail & Related papers (2024-02-05T06:52:53Z) - Optimizing Active Learning for Low Annotation Budgets [6.753808772846254]
In deep learning, active learning is usually implemented as an iterative process in which successive deep models are updated via fine tuning.
We tackle this issue by using an approach inspired by transfer learning.
We introduce a novel acquisition function which exploits the iterative nature of AL process to select samples in a more robust fashion.
arXiv Detail & Related papers (2022-01-18T18:53:10Z) - DapStep: Deep Assignee Prediction for Stack Trace Error rePresentation [61.99379022383108]
We propose new deep learning models to solve the bug triage problem.
The models are based on a bidirectional recurrent neural network with attention and on a convolutional neural network.
To improve the quality of ranking, we propose using additional information from version control system annotations.
arXiv Detail & Related papers (2022-01-14T00:16:57Z) - Few-shot Learning for Topic Modeling [39.56814839510978]
We propose a neural network-based few-shot learning method that can learn a topic model from just a few documents.
We demonstrate that the proposed method achieves better perplexity than existing methods using three real-world text document sets.
arXiv Detail & Related papers (2021-04-19T01:56:48Z) - Pre-training for Abstractive Document Summarization by Reinstating
Source Text [105.77348528847337]
This paper presents three pre-training objectives which allow us to pre-train a Seq2Seq based abstractive summarization model on unlabeled text.
Experiments on two benchmark summarization datasets show that all three objectives can improve performance upon baselines.
arXiv Detail & Related papers (2020-04-04T05:06:26Z) - Document Ranking with a Pretrained Sequence-to-Sequence Model [56.44269917346376]
We show how a sequence-to-sequence model can be trained to generate relevance labels as "target words"
Our approach significantly outperforms an encoder-only model in a data-poor regime.
arXiv Detail & Related papers (2020-03-14T22:29:50Z) - Pre-training Tasks for Embedding-based Large-scale Retrieval [68.01167604281578]
We consider the large-scale query-document retrieval problem.
Given a query (e.g., a question), return the set of relevant documents from a large document corpus.
We show that the key ingredient of learning a strong embedding-based Transformer model is the set of pre-training tasks.
arXiv Detail & Related papers (2020-02-10T16:44:00Z) - AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses [97.50616524350123]
We build dialogue models that are dynamically aware of what utterances or tokens are dull without any feature-engineering.
The first model, MinAvgOut, directly maximizes the diversity score through the output distributions of each batch.
The second model, Label Fine-Tuning (LFT), prepends to the source sequence a label continuously scaled by the diversity score to control the diversity level.
The third model, RL, adopts Reinforcement Learning and treats the diversity score as a reward signal.
arXiv Detail & Related papers (2020-01-15T18:32:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.