Which Tricks Are Important for Learning to Rank?
- URL: http://arxiv.org/abs/2204.01500v2
- Date: Fri, 6 Oct 2023 18:09:44 GMT
- Title: Which Tricks Are Important for Learning to Rank?
- Authors: Ivan Lyzhin, Aleksei Ustimenko, Andrey Gulin, Liudmila Prokhorenkova
- Abstract summary: State-of-the-art learning-to-rank methods are based on gradient-boosted decision trees (GBDT)
In this paper, we thoroughly analyze several GBDT-based ranking algorithms in a unified setup.
As a result, we gain insights into learning-to-rank techniques and obtain a new state-of-the-art algorithm.
- Score: 32.38701971636441
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Nowadays, state-of-the-art learning-to-rank methods are based on
gradient-boosted decision trees (GBDT). The most well-known algorithm is
LambdaMART which was proposed more than a decade ago. Recently, several other
GBDT-based ranking algorithms were proposed. In this paper, we thoroughly
analyze these methods in a unified setup. In particular, we address the
following questions. Is direct optimization of a smoothed ranking loss
preferable over optimizing a convex surrogate? How to properly construct and
smooth surrogate ranking losses? To address these questions, we compare
LambdaMART with YetiRank and StochasticRank methods and their modifications. We
also propose a simple improvement of the YetiRank approach that allows for
optimizing specific ranking loss functions. As a result, we gain insights into
learning-to-rank techniques and obtain a new state-of-the-art algorithm.
Related papers
- Newton Losses: Using Curvature Information for Learning with Differentiable Algorithms [80.37846867546517]
We show how to train eight different neural networks with custom objectives.
We exploit their second-order information via their empirical Fisherssian matrices.
We apply Loss Lossiable algorithms to achieve significant improvements for less differentiable algorithms.
arXiv Detail & Related papers (2024-10-24T18:02:11Z) - Learning the Positions in CountSketch [49.57951567374372]
We consider sketching algorithms which first compress data by multiplication with a random sketch matrix, and then apply the sketch to quickly solve an optimization problem.
In this work, we propose the first learning-based algorithms that also optimize the locations of the non-zero entries.
arXiv Detail & Related papers (2023-06-11T07:28:35Z) - Optimizing NOTEARS Objectives via Topological Swaps [41.18829644248979]
In this paper, we develop an effective method for a set of candidate algorithms.
At the inner level, given a subject, we utilize off-the-the-art constraints.
The proposed method significantly improves the score of other algorithms.
arXiv Detail & Related papers (2023-05-26T21:49:37Z) - HARRIS: Hybrid Ranking and Regression Forests for Algorithm Selection [75.84584400866254]
We propose a new algorithm selector leveraging special forests, combining the strengths of both approaches while alleviating their weaknesses.
HARRIS' decisions are based on a forest model, whose trees are created based on optimized on a hybrid ranking and regression loss function.
arXiv Detail & Related papers (2022-10-31T14:06:11Z) - Optimizing Ranking Systems Online as Bandits [2.282430531490685]
We study and propose solutions for four challenges in optimizing ranking systems online.
The effectiveness is related to how fast the algorithm learns from interactions.
Second, the deployed algorithm should be safe, which means the algorithm only displays reasonable content to user requests.
Third, as users change their preferences constantly, the algorithm should handle the nonstationarity.
arXiv Detail & Related papers (2021-10-12T08:07:46Z) - An Accelerated Variance-Reduced Conditional Gradient Sliding Algorithm
for First-order and Zeroth-order Optimization [111.24899593052851]
Conditional gradient algorithm (also known as the Frank-Wolfe algorithm) has recently regained popularity in the machine learning community.
ARCS is the first zeroth-order conditional gradient sliding type algorithms solving convex problems in zeroth-order optimization.
In first-order optimization, the convergence results of ARCS substantially outperform previous algorithms in terms of the number of gradient query oracle.
arXiv Detail & Related papers (2021-09-18T07:08:11Z) - Machine Learning for Online Algorithm Selection under Censored Feedback [71.6879432974126]
In online algorithm selection (OAS), instances of an algorithmic problem class are presented to an agent one after another, and the agent has to quickly select a presumably best algorithm from a fixed set of candidate algorithms.
For decision problems such as satisfiability (SAT), quality typically refers to the algorithm's runtime.
In this work, we revisit multi-armed bandit algorithms for OAS and discuss their capability of dealing with the problem.
We adapt them towards runtime-oriented losses, allowing for partially censored data while keeping a space- and time-complexity independent of the time horizon.
arXiv Detail & Related papers (2021-09-13T18:10:52Z) - PiRank: Learning To Rank via Differentiable Sorting [85.28916333414145]
We propose PiRank, a new class of differentiable surrogates for ranking.
We show that PiRank exactly recovers the desired metrics in the limit of zero temperature.
arXiv Detail & Related papers (2020-12-12T05:07:36Z) - TopRank+: A Refinement of TopRank Algorithm [0.0]
A novel online learning algorithm was proposed based on topological sorting.
In this work, we utilize method of mixtures and expansions of certain implicit function to provide a tighter, iterated-log-like boundary for the inequalities.
arXiv Detail & Related papers (2020-01-21T15:44:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.