On the Linear Ordering Problem and the Rankability of Data
- URL: http://arxiv.org/abs/2104.05816v1
- Date: Mon, 12 Apr 2021 21:05:17 GMT
- Title: On the Linear Ordering Problem and the Rankability of Data
- Authors: Thomas R. Cameron, Sebastian Charmot, Jonad Pulaj
- Abstract summary: We use the degree of linearity to quantify what percentage of the data aligns with an optimal ranking.
In a sports context, this is analogous to the number of games that a ranking can correctly predict in hindsight.
We show that the optimal rankings computed via the LOP maximize the hindsight accuracy of a ranking.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In 2019, Anderson et al. proposed the concept of rankability, which refers to
a dataset's inherent ability to be meaningfully ranked. In this article, we
give an expository review of the linear ordering problem (LOP) and then use it
to analyze the rankability of data. Specifically, the degree of linearity is
used to quantify what percentage of the data aligns with an optimal ranking. In
a sports context, this is analogous to the number of games that a ranking can
correctly predict in hindsight. In fact, under the appropriate objective
function, we show that the optimal rankings computed via the LOP maximize the
hindsight accuracy of a ranking. Moreover, we develop a binary program to
compute the maximal Kendall tau ranking distance between two optimal rankings,
which can be used to measure the diversity among optimal rankings without
having to enumerate all optima. Finally, we provide several examples from the
world of sports and college rankings to illustrate these concepts and
demonstrate our results.
Related papers
- Soft Condorcet Optimization for Ranking of General Agents [44.90789674063613]
A common way to drive progress of AI models and agents is to compare their performance on standardized benchmarks.
In this paper, we describe a novel ranking scheme inspired by social choice frameworks, called Soft Condorcet Optimization (SCO)
SCO rankings are on average 0 to 0.043 away from the optimal ranking in normalized Kendall-tau distance across 865 preference profiles from the PrefLib open ranking archive.
arXiv Detail & Related papers (2024-10-31T18:17:39Z) - Adaptive Neural Ranking Framework: Toward Maximized Business Goal for
Cascade Ranking Systems [33.46891569350896]
Cascade ranking is widely used for large-scale top-k selection problems in online advertising and recommendation systems.
Previous works on learning-to-rank usually focus on letting the model learn the complete order or top-k order.
We name this method as Adaptive Neural Ranking Framework (abbreviated as ARF)
arXiv Detail & Related papers (2023-10-16T14:43:02Z) - Parameterized Aspects of Distinct Kemeny Rank Aggregation [4.792851066169871]
We study the problem of computing all distinct Kemeny rankings under the lens of parameterized complexity.
We find that any desirable number of Kemeny rankings can also be found without substantial increase in running time.
arXiv Detail & Related papers (2023-09-07T06:58:19Z) - Bipartite Ranking Fairness through a Model Agnostic Ordering Adjustment [54.179859639868646]
We propose a model agnostic post-processing framework xOrder for achieving fairness in bipartite ranking.
xOrder is compatible with various classification models and ranking fairness metrics, including supervised and unsupervised fairness metrics.
We evaluate our proposed algorithm on four benchmark data sets and two real-world patient electronic health record repositories.
arXiv Detail & Related papers (2023-07-27T07:42:44Z) - Optimizing Partial Area Under the Top-k Curve: Theory and Practice [151.5072746015253]
We develop a novel metric named partial Area Under the top-k Curve (AUTKC)
AUTKC has a better discrimination ability, and its Bayes optimal score function could give a correct top-K ranking with respect to the conditional probability.
We present an empirical surrogate risk minimization framework to optimize the proposed metric.
arXiv Detail & Related papers (2022-09-03T11:09:13Z) - GNNRank: Learning Global Rankings from Pairwise Comparisons via Directed
Graph Neural Networks [68.61934077627085]
We introduce GNNRank, a modeling framework compatible with any GNN capable of learning digraph embeddings.
We show that our methods attain competitive and often superior performance compared with existing approaches.
arXiv Detail & Related papers (2022-02-01T04:19:50Z) - Data Driven and Visualization based Strategization for University Rank
Improvement using Decision Trees [1.933681537640272]
We present a novel idea of classifying the rankings data using Decision Tree (DT) based algorithms and retrieve decision paths for rank improvement using data visualization techniques.
The proposed methodology can aid HEIs to quantitatively asses the scope of improvement, adumbrate a fine-grained long-term action plan and prepare a suitable road-map.
arXiv Detail & Related papers (2021-10-18T06:41:45Z) - PiRank: Learning To Rank via Differentiable Sorting [85.28916333414145]
We propose PiRank, a new class of differentiable surrogates for ranking.
We show that PiRank exactly recovers the desired metrics in the limit of zero temperature.
arXiv Detail & Related papers (2020-12-12T05:07:36Z) - Towards Model-Agnostic Post-Hoc Adjustment for Balancing Ranking
Fairness and Algorithm Utility [54.179859639868646]
Bipartite ranking aims to learn a scoring function that ranks positive individuals higher than negative ones from labeled data.
There have been rising concerns on whether the learned scoring function can cause systematic disparity across different protected groups.
We propose a model post-processing framework for balancing them in the bipartite ranking scenario.
arXiv Detail & Related papers (2020-06-15T10:08:39Z) - How Reliable are University Rankings? [0.7646713951724009]
We take a fresh look at this ranking scheme using the public College dataset.
We show in multiple ways that this ranking scheme is not reliable and cannot be trusted as authoritative.
We conclude by making the case that all data and methods used for rankings should be made open for validation and repeatability.
arXiv Detail & Related papers (2020-04-20T01:00:59Z) - SetRank: A Setwise Bayesian Approach for Collaborative Ranking from
Implicit Feedback [50.13745601531148]
We propose a novel setwise Bayesian approach for collaborative ranking, namely SetRank, to accommodate the characteristics of implicit feedback in recommender system.
Specifically, SetRank aims at maximizing the posterior probability of novel setwise preference comparisons.
We also present the theoretical analysis of SetRank to show that the bound of excess risk can be proportional to $sqrtM/N$.
arXiv Detail & Related papers (2020-02-23T06:40:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.