Learning-to-Rank at the Speed of Sampling: Plackett-Luce Gradient
Estimation With Minimal Computational Complexity
- URL: http://arxiv.org/abs/2204.10872v1
- Date: Fri, 22 Apr 2022 18:01:33 GMT
- Title: Learning-to-Rank at the Speed of Sampling: Plackett-Luce Gradient
Estimation With Minimal Computational Complexity
- Authors: Harrie Oosterhuis
- Abstract summary: We introduce the novel PL-Rank-3 algorithm that performs unbiased gradient estimation with a computational complexity comparable to the best sorting algorithms.
Our experimental results indicate large gains in the time required for optimization, without any loss in performance.
- Score: 13.579420996461439
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Plackett-Luce gradient estimation enables the optimization of stochastic
ranking models within feasible time constraints through sampling techniques.
Unfortunately, the computational complexity of existing methods does not scale
well with the length of the rankings, i.e. the ranking cutoff, nor with the
item collection size. In this paper, we introduce the novel PL-Rank-3 algorithm
that performs unbiased gradient estimation with a computational complexity
comparable to the best sorting algorithms. As a result, our novel
learning-to-rank method is applicable in any scenario where standard sorting is
feasible in reasonable time. Our experimental results indicate large gains in
the time required for optimization, without any loss in performance. For the
field, our contribution could potentially allow state-of-the-art
learning-to-rank methods to be applied to much larger scales than previously
feasible.
Related papers
- Ordering for Non-Replacement SGD [7.11967773739707]
We seek to find an ordering that can improve the convergence rates for the non-replacement form of the algorithm.
We develop optimal orderings for constant and decreasing step sizes for strongly convex and convex functions.
In addition, we are able to combine the ordering with mini-batch and further apply it to more complex neural networks.
arXiv Detail & Related papers (2023-06-28T00:46:58Z) - Optimal Algorithms for Stochastic Complementary Composite Minimization [55.26935605535377]
Inspired by regularization techniques in statistics and machine learning, we study complementary composite minimization.
We provide novel excess risk bounds, both in expectation and with high probability.
Our algorithms are nearly optimal, which we prove via novel lower complexity bounds for this class of problems.
arXiv Detail & Related papers (2022-11-03T12:40:24Z) - Fast Offline Policy Optimization for Large Scale Recommendation [74.78213147859236]
We derive an approximation of these policy learning algorithms that scale logarithmically with the catalogue size.
Our contribution is based upon combining three novel ideas.
Our estimator is an order of magnitude faster than naive approaches yet produces equally good policies.
arXiv Detail & Related papers (2022-08-08T11:54:11Z) - Local policy search with Bayesian optimization [73.0364959221845]
Reinforcement learning aims to find an optimal policy by interaction with an environment.
Policy gradients for local search are often obtained from random perturbations.
We develop an algorithm utilizing a probabilistic model of the objective function and its gradient.
arXiv Detail & Related papers (2021-06-22T16:07:02Z) - Estimating leverage scores via rank revealing methods and randomization [50.591267188664666]
We study algorithms for estimating the statistical leverage scores of rectangular dense or sparse matrices of arbitrary rank.
Our approach is based on combining rank revealing methods with compositions of dense and sparse randomized dimensionality reduction transforms.
arXiv Detail & Related papers (2021-05-23T19:21:55Z) - PiRank: Learning To Rank via Differentiable Sorting [85.28916333414145]
We propose PiRank, a new class of differentiable surrogates for ranking.
We show that PiRank exactly recovers the desired metrics in the limit of zero temperature.
arXiv Detail & Related papers (2020-12-12T05:07:36Z) - Effective Proximal Methods for Non-convex Non-smooth Regularized
Learning [27.775096437736973]
We show that the independent sampling scheme tends to improve performance of the commonly-used uniform sampling scheme.
Our new analysis also derives a speed for the sampling than best one available so far.
arXiv Detail & Related papers (2020-09-14T16:41:32Z) - Beyond Worst-Case Analysis in Stochastic Approximation: Moment
Estimation Improves Instance Complexity [58.70807593332932]
We study oracle complexity of gradient based methods for approximation problems.
We focus on instance-dependent complexity instead of worst case complexity.
Our proposed algorithm and its analysis provide a theoretical justification for the success of moment estimation.
arXiv Detail & Related papers (2020-06-08T09:25:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.