Improving Sequential Recommendation Models with an Enhanced Loss
Function
- URL: http://arxiv.org/abs/2301.00979v2
- Date: Mon, 22 May 2023 02:48:16 GMT
- Title: Improving Sequential Recommendation Models with an Enhanced Loss
Function
- Authors: Fangyu Li and Shenbao Yu and Feng Zeng and Fang Yang
- Abstract summary: We develop an improved loss function for sequential recommendation models.
We conduct experiments on two influential open-source libraries.
We reproduce the results of the BERT4Rec model on the Beauty dataset.
- Score: 9.573139673704766
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There has been a growing interest in benchmarking sequential recommendation
models and reproducing/improving existing models. For example, Rendle et al.
improved matrix factorization models by tuning their parameters and
hyperparameters. Petrov and Macdonald developed a more efficient and effective
implementation of BERT4Rec, which resolved inconsistencies in performance
comparison between BERT4Rec and SASRec in previous works. In particular,
BERT4Rec and SASRec share a similar network structure, with the main difference
lying in their training objective/loss function. Therefore, we analyzed the
advantages and disadvantages of commonly used loss functions in sequential
recommendation and proposed an improved loss function that leverages their
strengths. We conduct extensive experiments on two influential open-source
libraries, and the results demonstrate that our improved loss function
significantly enhances the performance of GRU4Rec, SASRec, SR-GNN, and S3Rec
models, improving their benchmarks significantly. Furthermore, the improved
SASRec benchmark outperforms BERT4Rec on the ML-1M and Beauty datasets and
achieves similar results to BERT4Rec on the ML-20M and Steam datasets. We also
reproduce the results of the BERT4Rec model on the Beauty dataset. Finally, we
provide a comprehensive explanation of the effectiveness of our improved loss
function through experiments. Our code is publicly available at
https://github.com/Li-fAngyU/sequential_rec.
Related papers
- Aligning GPTRec with Beyond-Accuracy Goals with Reinforcement Learning [67.71952251641545]
GPTRec is an alternative to the Top-K model for item-by-item recommendations.
We show that GPTRec offers a better tradeoff between accuracy and secondary metrics than classic greedy re-ranking techniques.
Our experiments on two datasets show that GPTRec's Next-K generation approach offers a better tradeoff between accuracy and secondary metrics than classic greedy re-ranking techniques.
arXiv Detail & Related papers (2024-03-07T19:47:48Z) - Efficient Grammatical Error Correction Via Multi-Task Training and
Optimized Training Schedule [55.08778142798106]
We propose auxiliary tasks that exploit the alignment between the original and corrected sentences.
We formulate each task as a sequence-to-sequence problem and perform multi-task training.
We find that the order of datasets used for training and even individual instances within a dataset may have important effects on the final performance.
arXiv Detail & Related papers (2023-11-20T14:50:12Z) - Turning Dross Into Gold Loss: is BERT4Rec really better than SASRec? [1.223779595809275]
Two state-of-the-art baselines are Transformer-based models SASRec and BERT4Rec.
In most of the publications, BERT4Rec achieves better performance than SASRec.
We show that SASRec could be effectively trained with negative sampling and still outperform BERT4Rec, but the number of negative examples should be much larger than one.
arXiv Detail & Related papers (2023-09-14T11:07:10Z) - gSASRec: Reducing Overconfidence in Sequential Recommendation Trained
with Negative Sampling [67.71952251641545]
We show that models trained with negative sampling tend to overestimate the probabilities of positive interactions.
We propose a novel Generalised Binary Cross-Entropy Loss function (gBCE) and theoretically prove that it can mitigate overconfidence.
We show through detailed experiments on three datasets that gSASRec does not exhibit the overconfidence problem.
arXiv Detail & Related papers (2023-08-14T14:56:40Z) - Scaling Relationship on Learning Mathematical Reasoning with Large
Language Models [75.29595679428105]
We investigate how the pre-training loss, supervised data amount, and augmented data amount influence the reasoning performances of a supervised LLM.
We find that rejection samples from multiple models push LLaMA-7B to an accuracy of 49.3% on GSM8K which outperforms the supervised fine-tuning (SFT) accuracy of 35.9% significantly.
arXiv Detail & Related papers (2023-08-03T15:34:01Z) - A Systematic Review and Replicability Study of BERT4Rec for Sequential
Recommendation [91.02268704681124]
BERT4Rec is an effective model for sequential recommendation based on the Transformer architecture.
We show that BERT4Rec results are not consistent within these publications.
We propose our own implementation of BERT4Rec based on the Hugging Face Transformers library.
arXiv Detail & Related papers (2022-07-15T14:09:04Z) - Using calibrator to improve robustness in Machine Reading Comprehension [18.844528744164876]
We propose a method to improve the robustness by using a calibrator as the post-hoc reranker.
Experimental results on adversarial datasets show that our model can achieve performance improvement by more than 10%.
arXiv Detail & Related papers (2022-02-24T02:16:42Z) - SPLADE v2: Sparse Lexical and Expansion Model for Information Retrieval [11.38022203865326]
SPLADE model provides highly sparse representations and competitive results with respect to state-of-the-art dense and sparse approaches.
We modify the pooling mechanism, benchmark a model solely based on document expansion, and introduce models trained with distillation.
Overall, SPLADE is considerably improved with more than $9$% gains on NDCG@10 on TREC DL 2019, leading to state-of-the-art results on the BEIR benchmark.
arXiv Detail & Related papers (2021-09-21T10:43:42Z) - S^3-Rec: Self-Supervised Learning for Sequential Recommendation with
Mutual Information Maximization [104.87483578308526]
We propose the model S3-Rec, which stands for Self-Supervised learning for Sequential Recommendation.
For our task, we devise four auxiliary self-supervised objectives to learn the correlations among attribute, item, subsequence, and sequence.
Extensive experiments conducted on six real-world datasets demonstrate the superiority of our proposed method over existing state-of-the-art methods.
arXiv Detail & Related papers (2020-08-18T11:44:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.