Ranking In Generalized Linear Bandits
- URL: http://arxiv.org/abs/2207.00109v2
- Date: Mon, 1 Jan 2024 21:27:44 GMT
- Title: Ranking In Generalized Linear Bandits
- Authors: Amitis Shidani, George Deligiannidis, Arnaud Doucet
- Abstract summary: We study the ranking problem in generalized linear bandits.
In recommendation systems, displaying an ordered list of the most attractive items is not always optimal.
- Score: 38.567816347428774
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study the ranking problem in generalized linear bandits. At each time, the
learning agent selects an ordered list of items and observes stochastic
outcomes. In recommendation systems, displaying an ordered list of the most
attractive items is not always optimal as both position and item dependencies
result in a complex reward function. A very naive example is the lack of
diversity when all the most attractive items are from the same category. We
model the position and item dependencies in the ordered list and design UCB and
Thompson Sampling type algorithms for this problem. Our work generalizes
existing studies in several directions, including position dependencies where
position discount is a particular case, and connecting the ranking problem to
graph theory.
Related papers
- A consensus set for the aggregation of partial rankings: the case of the Optimal Set of Bucket Orders Problem [46.752018687842344]
In rank aggregation problems (RAP), the solution is usually a consensus ranking that generalizes a set of input orderings.
We propose to provide, as the solution to an RAP, a set of rankings to better explain the preferences expressed in the input orderings.
arXiv Detail & Related papers (2025-02-19T14:32:16Z) - Found in the Middle: Permutation Self-Consistency Improves Listwise Ranking in Large Language Models [63.714662435555674]
Large language models (LLMs) exhibit positional bias in how they use context.
We propose permutation self-consistency, a form of self-consistency over ranking list outputs of black-box LLMs.
Our approach improves scores from conventional inference by up to 7-18% for GPT-3.5 and 8-16% for LLaMA v2 (70B)
arXiv Detail & Related papers (2023-10-11T17:59:02Z) - Learning to Select and Rank from Choice-Based Feedback: A Simple Nested Approach [10.293894471295205]
We study a ranking and selection problem of learning from choice-based feedback with dynamic assortments.
We present novel and simple algorithms for both learning goals.
arXiv Detail & Related papers (2023-07-13T05:05:30Z) - Learning List-Level Domain-Invariant Representations for Ranking [59.3544317373004]
We propose list-level alignment -- learning domain-invariant representations at the higher level of lists.
The benefits are twofold: it leads to the first domain adaptation generalization bound for ranking, in turn providing theoretical support for the proposed method.
arXiv Detail & Related papers (2022-12-21T04:49:55Z) - Online Learning of Optimally Diverse Rankings [63.62764375279861]
We propose an algorithm that efficiently learns the optimal list based on users' feedback only.
We show that after $T$ queries, the regret of LDR scales as $O((N-L)log(T))$ where $N$ is the number of all items.
arXiv Detail & Related papers (2021-09-13T12:13:20Z) - Aggregating Incomplete and Noisy Rankings [13.267203883254087]
We consider the problem of learning the true ordering of a set of alternatives from largely incomplete and noisy rankings.
Our selective Mallows model outputs a noisy ranking on any given subset of alternatives.
We show how to efficiently compute the maximum likelihood complete ranking from selective Mallows rankings.
arXiv Detail & Related papers (2020-11-02T08:18:33Z) - Distance-based Positive and Unlabeled Learning for Ranking [13.339237388350043]
Learning to rank is a problem of general interest.
We show that learning to rank via combining representations using an integer linear program is effective when the supervision is as light as "these few items are similar to your item of interest"
arXiv Detail & Related papers (2020-05-20T01:53:58Z) - SetRank: A Setwise Bayesian Approach for Collaborative Ranking from
Implicit Feedback [50.13745601531148]
We propose a novel setwise Bayesian approach for collaborative ranking, namely SetRank, to accommodate the characteristics of implicit feedback in recommender system.
Specifically, SetRank aims at maximizing the posterior probability of novel setwise preference comparisons.
We also present the theoretical analysis of SetRank to show that the bound of excess risk can be proportional to $sqrtM/N$.
arXiv Detail & Related papers (2020-02-23T06:40:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.