Determinantal Point Process Likelihoods for Sequential Recommendation
- URL: http://arxiv.org/abs/2204.11562v1
- Date: Mon, 25 Apr 2022 11:20:10 GMT
- Title: Determinantal Point Process Likelihoods for Sequential Recommendation
- Authors: Yuli Liu, Christian Walder, Lexing Xie
- Abstract summary: We propose two new loss functions based on the Determinantal Point Process (DPP) likelihood, that can be adaptively applied to estimate the subsequent item or items.
Experimental results using the proposed loss functions on three real-world datasets show marked improvements over state-of-the-art sequential recommendation methods in both quality and diversity metrics.
- Score: 12.206748373325972
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Sequential recommendation is a popular task in academic research and close to
real-world application scenarios, where the goal is to predict the next
action(s) of the user based on his/her previous sequence of actions. In the
training process of recommender systems, the loss function plays an essential
role in guiding the optimization of recommendation models to generate accurate
suggestions for users. However, most existing sequential recommendation
techniques focus on designing algorithms or neural network architectures, and
few efforts have been made to tailor loss functions that fit naturally into the
practical application scenario of sequential recommender systems.
Ranking-based losses, such as cross-entropy and Bayesian Personalized Ranking
(BPR) are widely used in the sequential recommendation area. We argue that such
objective functions suffer from two inherent drawbacks: i) the dependencies
among elements of a sequence are overlooked in these loss formulations; ii)
instead of balancing accuracy (quality) and diversity, only generating accurate
results has been over emphasized. We therefore propose two new loss functions
based on the Determinantal Point Process (DPP) likelihood, that can be
adaptively applied to estimate the subsequent item or items. The
DPP-distributed item set captures natural dependencies among temporal actions,
and a quality vs. diversity decomposition of the DPP kernel pushes us to go
beyond accuracy-oriented loss functions. Experimental results using the
proposed loss functions on three real-world datasets show marked improvements
over state-of-the-art sequential recommendation methods in both quality and
diversity metrics.
Related papers
- Generalizing Bayesian Optimization with Decision-theoretic Entropies [102.82152945324381]
We consider a generalization of Shannon entropy from work in statistical decision theory.
We first show that special cases of this entropy lead to popular acquisition functions used in BO procedures.
We then show how alternative choices for the loss yield a flexible family of acquisition functions.
arXiv Detail & Related papers (2022-10-04T04:43:58Z) - SPR:Supervised Personalized Ranking Based on Prior Knowledge for
Recommendation [6.407166061614783]
We propose a novel loss function named Supervised Personalized Ranking (SPR) Based on Prior Knowledge.
Unlike BPR, instead of constructing user, positive item, negative item> triples, the proposed SPR constructs user, similar user, positive item, negative item> quadruples.
arXiv Detail & Related papers (2022-07-07T10:00:54Z) - Recommendation Systems with Distribution-Free Reliability Guarantees [83.80644194980042]
We show how to return a set of items rigorously guaranteed to contain mostly good items.
Our procedure endows any ranking model with rigorous finite-sample control of the false discovery rate.
We evaluate our methods on the Yahoo! Learning to Rank and MSMarco datasets.
arXiv Detail & Related papers (2022-07-04T17:49:25Z) - Gleo-Det: Deep Convolution Feature-Guided Detector with Local Entropy
Optimization for Salient Points [5.955667705173262]
We propose to achieve fine constraint based on the requirement of repeatability while coarse constraint with guidance of deep convolution features.
With the guidance of convolution features, we define the cost function from both positive and negative sides.
arXiv Detail & Related papers (2022-04-27T12:40:21Z) - Robust and Adaptive Temporal-Difference Learning Using An Ensemble of
Gaussian Processes [70.80716221080118]
The paper takes a generative perspective on policy evaluation via temporal-difference (TD) learning.
The OS-GPTD approach is developed to estimate the value function for a given policy by observing a sequence of state-reward pairs.
To alleviate the limited expressiveness associated with a single fixed kernel, a weighted ensemble (E) of GP priors is employed to yield an alternative scheme.
arXiv Detail & Related papers (2021-12-01T23:15:09Z) - Dynamic Iterative Refinement for Efficient 3D Hand Pose Estimation [87.54604263202941]
We propose a tiny deep neural network of which partial layers are iteratively exploited for refining its previous estimations.
We employ learned gating criteria to decide whether to exit from the weight-sharing loop, allowing per-sample adaptation in our model.
Our method consistently outperforms state-of-the-art 2D/3D hand pose estimation approaches in terms of both accuracy and efficiency for widely used benchmarks.
arXiv Detail & Related papers (2021-11-11T23:31:34Z) - D2RLIR : an improved and diversified ranking function in interactive
recommendation systems based on deep reinforcement learning [0.3058685580689604]
This paper proposes a deep reinforcement learning based recommendation system by utilizing Actor-Critic architecture.
The proposed model is able to generate a diverse while relevance recommendation list based on the user's preferences.
arXiv Detail & Related papers (2021-10-28T13:11:29Z) - The Perils of Learning Before Optimizing [16.97597806975415]
We show how prediction models can be learned end-to-end by differentiating through the optimization task.
We show that the performance gap between a two-stage and end-to-end approach is closely related to the emphprice of correlation concept in optimization.
arXiv Detail & Related papers (2021-06-18T20:43:47Z) - Learning MDPs from Features: Predict-Then-Optimize for Sequential
Decision Problems by Reinforcement Learning [52.74071439183113]
We study the predict-then-optimize framework in the context of sequential decision problems (formulated as MDPs) solved via reinforcement learning.
Two significant computational challenges arise in applying decision-focused learning to MDPs.
arXiv Detail & Related papers (2021-06-06T23:53:31Z) - Optimizing for the Future in Non-Stationary MDPs [52.373873622008944]
We present a policy gradient algorithm that maximizes a forecast of future performance.
We show that our algorithm, called Prognosticator, is more robust to non-stationarity than two online adaptation techniques.
arXiv Detail & Related papers (2020-05-17T03:41:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.