The Performance of the MLE in the Bradley-Terry-Luce Model in
$\ell_{\infty}$-Loss and under General Graph Topologies
- URL: http://arxiv.org/abs/2110.10825v1
- Date: Wed, 20 Oct 2021 23:46:35 GMT
- Title: The Performance of the MLE in the Bradley-Terry-Luce Model in
$\ell_{\infty}$-Loss and under General Graph Topologies
- Authors: Wanshan Li, Shamindra Shrotriya, Alessandro Rinaldo
- Abstract summary: We derive novel, general upper bounds on the $ell_infty$ estimation error of the Bradley-Terry-Luce model.
We demonstrate that the derived bounds perform well and in some cases are sharper compared to known results.
- Score: 76.61051540383494
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Bradley-Terry-Luce (BTL) model is a popular statistical approach for
estimating the global ranking of a collection of items of interest using
pairwise comparisons. To ensure accurate ranking, it is essential to obtain
precise estimates of the model parameters in the $\ell_{\infty}$-loss. The
difficulty of this task depends crucially on the topology of the pairwise
comparison graph over the given items. However, beyond very few well-studied
cases, such as the complete and Erd\"os-R\'enyi comparison graphs, little is
known about the performance of the maximum likelihood estimator (MLE) of the
BTL model parameters in the $\ell_{\infty}$-loss under more general graph
topologies. In this paper, we derive novel, general upper bounds on the
$\ell_{\infty}$ estimation error of the BTL MLE that depend explicitly on the
algebraic connectivity of the comparison graph, the maximal performance gap
across items and the sample complexity. We demonstrate that the derived bounds
perform well and in some cases are sharper compared to known results obtained
using different loss functions and more restricted assumptions and graph
topologies. We further provide minimax lower bounds under $\ell_{\infty}$-error
that nearly match the upper bounds over a class of sufficiently regular graph
topologies. Finally, we study the implications of our bounds for efficient
tournament design. We illustrate and discuss our findings through various
examples and simulations.
Related papers
- Minimax Hypothesis Testing for the Bradley-Terry-Luce Model [6.5990719141691825]
The Bradley-Terry-Luce (BTL) model is one of the most widely used models for ranking a collection of items or agents.
We propose a hypothesis test that determines whether a given pairwise comparison dataset, with $k$ comparisons per pair of agents, originates from an underlying BTL model.
arXiv Detail & Related papers (2024-10-10T20:28:05Z) - Top-$K$ ranking with a monotone adversary [19.871049853222132]
We consider the scenario where a comparison graph is randomly generated and the adversary is allowed to add arbitrary edges.
The statistician's goal is then to accurately identify the top-$K$ preferred items based on pairwise comparisons.
The main contribution of this paper is to develop a weighted maximum likelihood estimator (MLE) that achieves near-optimal sample complexity.
arXiv Detail & Related papers (2024-02-12T06:57:34Z) - Provably Efficient High-Dimensional Bandit Learning with Batched
Feedbacks [93.00280593719513]
We study high-dimensional multi-armed contextual bandits with batched feedback where the $T$ steps of online interactions are divided into $L$ batches.
In specific, each batch collects data according to a policy that depends on previous batches and the rewards are revealed only at the end of the batch.
Our algorithm achieves regret bounds comparable to those in fully sequential setting with only $mathcalO( log T)$ batches.
arXiv Detail & Related papers (2023-11-22T06:06:54Z) - Degree Heterogeneity in Higher-Order Networks: Inference in the Hypergraph $\boldsymbolβ$-Model [4.540236408836132]
We study the hypergraph $boldsymbolbeta$-model with multiple layers, which allows for hyperedges of different sizes across the layers.
We derive the rates of convergence of maximum likelihood (ML) estimates and establish their minimax rate optimality.
We also consider the goodness-of-fit problem in the hypergraph $boldsymbolbeta$-model.
arXiv Detail & Related papers (2023-07-06T07:23:06Z) - Principled Reinforcement Learning with Human Feedback from Pairwise or
$K$-wise Comparisons [79.98542868281473]
We provide a theoretical framework for Reinforcement Learning with Human Feedback (RLHF)
We show that when training a policy based on the learned reward model, MLE fails while a pessimistic MLE provides policies with improved performance under certain coverage assumptions.
arXiv Detail & Related papers (2023-01-26T18:07:21Z) - Uncertainty Quantification of MLE for Entity Ranking with Covariates [3.2839905453386162]
This paper concerns with statistical estimation and inference for the ranking problems based on pairwise comparisons.
We propose a novel model, Co-Assisted Ranking Estimation (CARE) model, that extends the well-known Bradley-Terry-Luce (BTL) model.
We derive the maximum likelihood estimator of $alpha_i*_i=1n$ and $beta*$ under a sparse comparison graph.
We validate our theoretical results through large-scale numerical studies and an application to the mutual fund stock holding dataset.
arXiv Detail & Related papers (2022-12-20T02:28:27Z) - Multi-task Learning of Order-Consistent Causal Graphs [59.9575145128345]
We consider the problem of discovering $K related Gaussian acyclic graphs (DAGs)
Under multi-task learning setting, we propose a $l_1/l$-regularized maximum likelihood estimator (MLE) for learning $K$ linear structural equation models.
We theoretically show that the joint estimator, by leveraging data across related tasks, can achieve a better sample complexity for recovering the causal order.
arXiv Detail & Related papers (2021-11-03T22:10:18Z) - Sample-efficient L0-L2 constrained structure learning of sparse Ising
models [3.056751497358646]
We consider the problem of learning the underlying graph of a sparse Ising model with $p$ nodes from $n$ i.i.d. samples.
We leverage the cardinality constraint L0 norm, which is known to properly induce sparsity, and further combine it with an L2 norm to better model the non-zero coefficients.
arXiv Detail & Related papers (2020-12-03T07:52:20Z) - Model-Based Multi-Agent RL in Zero-Sum Markov Games with Near-Optimal
Sample Complexity [67.02490430380415]
We show that model-based MARL achieves a sample complexity of $tilde O(|S||B|(gamma)-3epsilon-2)$ for finding the Nash equilibrium (NE) value up to some $epsilon$ error.
We also show that such a sample bound is minimax-optimal (up to logarithmic factors) if the algorithm is reward-agnostic, where the algorithm queries state transition samples without reward knowledge.
arXiv Detail & Related papers (2020-07-15T03:25:24Z) - Ranking a set of objects: a graph based least-square approach [70.7866286425868]
We consider the problem of ranking $N$ objects starting from a set of noisy pairwise comparisons provided by a crowd of equal workers.
We propose a class of non-adaptive ranking algorithms that rely on a least-squares intrinsic optimization criterion for the estimation of qualities.
arXiv Detail & Related papers (2020-02-26T16:19:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.