Ranking In Generalized Linear Bandits
- URL: http://arxiv.org/abs/2207.00109v2
- Date: Mon, 1 Jan 2024 21:27:44 GMT
- Title: Ranking In Generalized Linear Bandits
- Authors: Amitis Shidani, George Deligiannidis, Arnaud Doucet
- Abstract summary: We study the ranking problem in generalized linear bandits.
In recommendation systems, displaying an ordered list of the most attractive items is not always optimal.
- Score: 38.567816347428774
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study the ranking problem in generalized linear bandits. At each time, the
learning agent selects an ordered list of items and observes stochastic
outcomes. In recommendation systems, displaying an ordered list of the most
attractive items is not always optimal as both position and item dependencies
result in a complex reward function. A very naive example is the lack of
diversity when all the most attractive items are from the same category. We
model the position and item dependencies in the ordered list and design UCB and
Thompson Sampling type algorithms for this problem. Our work generalizes
existing studies in several directions, including position dependencies where
position discount is a particular case, and connecting the ranking problem to
graph theory.
Related papers
- Found in the Middle: Permutation Self-Consistency Improves Listwise Ranking in Large Language Models [63.714662435555674]
Large language models (LLMs) exhibit positional bias in how they use context.
We propose permutation self-consistency, a form of self-consistency over ranking list outputs of black-box LLMs.
Our approach improves scores from conventional inference by up to 7-18% for GPT-3.5 and 8-16% for LLaMA v2 (70B)
arXiv Detail & Related papers (2023-10-11T17:59:02Z) - Bipartite Ranking Fairness through a Model Agnostic Ordering Adjustment [54.179859639868646]
We propose a model agnostic post-processing framework xOrder for achieving fairness in bipartite ranking.
xOrder is compatible with various classification models and ranking fairness metrics, including supervised and unsupervised fairness metrics.
We evaluate our proposed algorithm on four benchmark data sets and two real-world patient electronic health record repositories.
arXiv Detail & Related papers (2023-07-27T07:42:44Z) - Learning List-Level Domain-Invariant Representations for Ranking [59.3544317373004]
We propose list-level alignment -- learning domain-invariant representations at the higher level of lists.
The benefits are twofold: it leads to the first domain adaptation generalization bound for ranking, in turn providing theoretical support for the proposed method.
arXiv Detail & Related papers (2022-12-21T04:49:55Z) - Unbiased Pairwise Learning to Rank in Recommender Systems [4.058828240864671]
Unbiased learning to rank algorithms are appealing candidates and have already been applied in many applications with single categorical labels.
We propose a novel unbiased LTR algorithm to tackle the challenges, which innovatively models position bias in the pairwise fashion.
Experiment results on public benchmark datasets and internal live traffic show the superior results of the proposed method for both categorical and continuous labels.
arXiv Detail & Related papers (2021-11-25T06:04:59Z) - Online Learning of Optimally Diverse Rankings [63.62764375279861]
We propose an algorithm that efficiently learns the optimal list based on users' feedback only.
We show that after $T$ queries, the regret of LDR scales as $O((N-L)log(T))$ where $N$ is the number of all items.
arXiv Detail & Related papers (2021-09-13T12:13:20Z) - Maximizing Store Revenues using Tabu Search for Floor Space Optimization [0.0]
We formulate the problem as a connected multi-choice knapsack problem with an additional global constraint.
We propose a tabu search based meta-heuristic that exploits the multiple special neighborhood structures.
arXiv Detail & Related papers (2020-11-04T22:42:54Z) - Aggregating Incomplete and Noisy Rankings [13.267203883254087]
We consider the problem of learning the true ordering of a set of alternatives from largely incomplete and noisy rankings.
Our selective Mallows model outputs a noisy ranking on any given subset of alternatives.
We show how to efficiently compute the maximum likelihood complete ranking from selective Mallows rankings.
arXiv Detail & Related papers (2020-11-02T08:18:33Z) - Distance-based Positive and Unlabeled Learning for Ranking [13.339237388350043]
Learning to rank is a problem of general interest.
We show that learning to rank via combining representations using an integer linear program is effective when the supervision is as light as "these few items are similar to your item of interest"
arXiv Detail & Related papers (2020-05-20T01:53:58Z) - SetRank: A Setwise Bayesian Approach for Collaborative Ranking from
Implicit Feedback [50.13745601531148]
We propose a novel setwise Bayesian approach for collaborative ranking, namely SetRank, to accommodate the characteristics of implicit feedback in recommender system.
Specifically, SetRank aims at maximizing the posterior probability of novel setwise preference comparisons.
We also present the theoretical analysis of SetRank to show that the bound of excess risk can be proportional to $sqrtM/N$.
arXiv Detail & Related papers (2020-02-23T06:40:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.