Determinantal Point Process Likelihoods for Sequential Recommendation
- URL: http://arxiv.org/abs/2204.11562v1
- Date: Mon, 25 Apr 2022 11:20:10 GMT
- Title: Determinantal Point Process Likelihoods for Sequential Recommendation
- Authors: Yuli Liu, Christian Walder, Lexing Xie
- Abstract summary: We propose two new loss functions based on the Determinantal Point Process (DPP) likelihood, that can be adaptively applied to estimate the subsequent item or items.
Experimental results using the proposed loss functions on three real-world datasets show marked improvements over state-of-the-art sequential recommendation methods in both quality and diversity metrics.
- Score: 12.206748373325972
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Sequential recommendation is a popular task in academic research and close to
real-world application scenarios, where the goal is to predict the next
action(s) of the user based on his/her previous sequence of actions. In the
training process of recommender systems, the loss function plays an essential
role in guiding the optimization of recommendation models to generate accurate
suggestions for users. However, most existing sequential recommendation
techniques focus on designing algorithms or neural network architectures, and
few efforts have been made to tailor loss functions that fit naturally into the
practical application scenario of sequential recommender systems.
Ranking-based losses, such as cross-entropy and Bayesian Personalized Ranking
(BPR) are widely used in the sequential recommendation area. We argue that such
objective functions suffer from two inherent drawbacks: i) the dependencies
among elements of a sequence are overlooked in these loss formulations; ii)
instead of balancing accuracy (quality) and diversity, only generating accurate
results has been over emphasized. We therefore propose two new loss functions
based on the Determinantal Point Process (DPP) likelihood, that can be
adaptively applied to estimate the subsequent item or items. The
DPP-distributed item set captures natural dependencies among temporal actions,
and a quality vs. diversity decomposition of the DPP kernel pushes us to go
beyond accuracy-oriented loss functions. Experimental results using the
proposed loss functions on three real-world datasets show marked improvements
over state-of-the-art sequential recommendation methods in both quality and
diversity metrics.
Related papers
- Uncertainty-Penalized Direct Preference Optimization [52.387088396044206]
We develop a pessimistic framework for DPO by introducing preference uncertainty penalization schemes.
The penalization serves as a correction to the loss which attenuates the loss gradient for uncertain samples.
We show improved overall performance compared to vanilla DPO, as well as better completions on prompts from high-uncertainty chosen/rejected responses.
arXiv Detail & Related papers (2024-10-26T14:24:37Z) - Exploiting Preferences in Loss Functions for Sequential Recommendation via Weak Transitivity [4.7894654945375175]
A choice of optimization objective is immensely pivotal in the design of a recommender system.
We propose a novel method that extends original objectives to explicitly leverage the different levels of preferences as relative orders between their scores.
arXiv Detail & Related papers (2024-08-01T06:55:19Z) - Learning k-Determinantal Point Processes for Personalized Ranking [13.677246792673564]
We present a new optimization criterion LkP based on set probability comparison for personalized ranking.
LkP is broadly applicable, and when applied to existing recommendation models it also yields strong performance improvements.
arXiv Detail & Related papers (2024-06-23T02:24:50Z) - Generalizing Bayesian Optimization with Decision-theoretic Entropies [102.82152945324381]
We consider a generalization of Shannon entropy from work in statistical decision theory.
We first show that special cases of this entropy lead to popular acquisition functions used in BO procedures.
We then show how alternative choices for the loss yield a flexible family of acquisition functions.
arXiv Detail & Related papers (2022-10-04T04:43:58Z) - Recommendation Systems with Distribution-Free Reliability Guarantees [83.80644194980042]
We show how to return a set of items rigorously guaranteed to contain mostly good items.
Our procedure endows any ranking model with rigorous finite-sample control of the false discovery rate.
We evaluate our methods on the Yahoo! Learning to Rank and MSMarco datasets.
arXiv Detail & Related papers (2022-07-04T17:49:25Z) - Gleo-Det: Deep Convolution Feature-Guided Detector with Local Entropy
Optimization for Salient Points [5.955667705173262]
We propose to achieve fine constraint based on the requirement of repeatability while coarse constraint with guidance of deep convolution features.
With the guidance of convolution features, we define the cost function from both positive and negative sides.
arXiv Detail & Related papers (2022-04-27T12:40:21Z) - Dynamic Iterative Refinement for Efficient 3D Hand Pose Estimation [87.54604263202941]
We propose a tiny deep neural network of which partial layers are iteratively exploited for refining its previous estimations.
We employ learned gating criteria to decide whether to exit from the weight-sharing loop, allowing per-sample adaptation in our model.
Our method consistently outperforms state-of-the-art 2D/3D hand pose estimation approaches in terms of both accuracy and efficiency for widely used benchmarks.
arXiv Detail & Related papers (2021-11-11T23:31:34Z) - The Perils of Learning Before Optimizing [16.97597806975415]
We show how prediction models can be learned end-to-end by differentiating through the optimization task.
We show that the performance gap between a two-stage and end-to-end approach is closely related to the emphprice of correlation concept in optimization.
arXiv Detail & Related papers (2021-06-18T20:43:47Z) - Learning MDPs from Features: Predict-Then-Optimize for Sequential
Decision Problems by Reinforcement Learning [52.74071439183113]
We study the predict-then-optimize framework in the context of sequential decision problems (formulated as MDPs) solved via reinforcement learning.
Two significant computational challenges arise in applying decision-focused learning to MDPs.
arXiv Detail & Related papers (2021-06-06T23:53:31Z) - Optimizing for the Future in Non-Stationary MDPs [52.373873622008944]
We present a policy gradient algorithm that maximizes a forecast of future performance.
We show that our algorithm, called Prognosticator, is more robust to non-stationarity than two online adaptation techniques.
arXiv Detail & Related papers (2020-05-17T03:41:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.