LLM4SBR: A Lightweight and Effective Framework for Integrating Large
Language Models in Session-based Recommendation
- URL: http://arxiv.org/abs/2402.13840v1
- Date: Wed, 21 Feb 2024 14:38:02 GMT
- Title: LLM4SBR: A Lightweight and Effective Framework for Integrating Large
Language Models in Session-based Recommendation
- Authors: Shutong Qiao, Chen Gao, Junhao Wen, Wei Zhou, Qun Luo, Peixuan Chen
and Yong Li
- Abstract summary: Traditional session-based recommendation (SBR) utilizes session behavior sequences from anonymous users for recommendation.
We propose the LLM Integration Framework for SBR (LLM4SBR) as a lightweight and plug-and-play framework.
We conducted experiments on two real-world datasets, and the results demonstrate that LLM4SBR significantly improves the performance of traditional SBR models.
- Score: 27.922143384779563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traditional session-based recommendation (SBR) utilizes session behavior
sequences from anonymous users for recommendation. Although this strategy is
highly efficient, it sacrifices the inherent semantic information of the items,
making it difficult for the model to understand the true intent of the session
and resulting in a lack of interpretability in the recommended results.
Recently, large language models (LLMs) have flourished across various domains,
offering a glimpse of hope in addressing the aforementioned challenges.
Inspired by the impact of LLMs, research exploring the integration of LLMs with
the Recommender system (RS) has surged like mushrooms after rain. However,
constrained by high time and space costs, as well as the brief and anonymous
nature of session data, the first LLM recommendation framework suitable for
industrial deployment has yet to emerge in the field of SBR. To address the
aforementioned challenges, we have proposed the LLM Integration Framework for
SBR (LLM4SBR). Serving as a lightweight and plug-and-play framework, LLM4SBR
adopts a two-step strategy. Firstly, we transform session data into a bimodal
form of text and behavior. In the first step, leveraging the inferential
capabilities of LLMs, we conduct inference on session text data from different
perspectives and design the component for auxiliary enhancement. In the second
step, the SBR model is trained on behavior data, aligning and averaging two
modal session representations from different perspectives. Finally, we fuse
session representations from different perspectives and modalities as the
ultimate session representation for recommendation. We conducted experiments on
two real-world datasets, and the results demonstrate that LLM4SBR significantly
improves the performance of traditional SBR models and is highly lightweight
and efficient, making it suitable for industrial deployment.
Related papers
- Align-SLM: Textless Spoken Language Models with Reinforcement Learning from AI Feedback [50.84142264245052]
This work introduces the Align-SLM framework to enhance the semantic understanding of textless Spoken Language Models (SLMs)
Our approach generates multiple speech continuations from a given prompt and uses semantic metrics to create preference data for Direct Preference Optimization (DPO)
We evaluate the framework using ZeroSpeech 2021 benchmarks for lexical and syntactic modeling, the spoken version of the StoryCloze dataset for semantic coherence, and other speech generation metrics, including the GPT4-o score and human evaluation.
arXiv Detail & Related papers (2024-11-04T06:07:53Z) - Item Cluster-aware Prompt Learning for Session-based Recommendation [36.93334485299296]
Session-based recommendation aims to capture user preferences by analyzing item sequences within individual sessions.
Most existing approaches focus mainly on intra-session item relationships, neglecting the connections between items across different sessions.
We propose the CLIP-SBR (Cluster-aware Item Prompt learning for Session-Based Recommendation) framework to address these challenges.
arXiv Detail & Related papers (2024-10-07T05:20:21Z) - Large Language Model Empowered Embedding Generator for Sequential Recommendation [57.49045064294086]
Large Language Model (LLM) has the potential to understand the semantic connections between items, regardless of their popularity.
We present LLMEmb, an innovative technique that harnesses LLM to create item embeddings that bolster the performance of Sequential Recommender Systems.
arXiv Detail & Related papers (2024-09-30T03:59:06Z) - DaRec: A Disentangled Alignment Framework for Large Language Model and Recommender System [83.34921966305804]
Large language models (LLMs) have demonstrated remarkable performance in recommender systems.
We propose a novel plug-and-play alignment framework for LLMs and collaborative models.
Our method is superior to existing state-of-the-art algorithms.
arXiv Detail & Related papers (2024-08-15T15:56:23Z) - Beyond Inter-Item Relations: Dynamic Adaption for Enhancing LLM-Based Sequential Recommendation [83.87767101732351]
Sequential recommender systems (SRS) predict the next items that users may prefer based on user historical interaction sequences.
Inspired by the rise of large language models (LLMs) in various AI applications, there is a surge of work on LLM-based SRS.
We propose DARec, a sequential recommendation model built on top of coarse-grained adaption for capturing inter-item relations.
arXiv Detail & Related papers (2024-08-14T10:03:40Z) - Understand What LLM Needs: Dual Preference Alignment for Retrieval-Augmented Generation [64.7982176398485]
Retrieval-augmented generation (RAG) has demonstrated effectiveness in mitigating the hallucination problem of large language models (LLMs)
We propose DPA-RAG, a universal framework designed to align diverse knowledge preferences within RAG systems.
arXiv Detail & Related papers (2024-06-26T18:26:53Z) - LLM4MSR: An LLM-Enhanced Paradigm for Multi-Scenario Recommendation [45.31960122494715]
We propose an efficient interpretable large language model (LLM)-enhanced paradigm LLM4MSR.
Specifically, we first leverage LLM to uncover multi-level knowledge including scenario correlations and users' cross-scenario interests.
Our experiments on KuaiSAR-small, KuaiSAR, and Amazon datasets validate two significant advantages of LLM4MSR.
arXiv Detail & Related papers (2024-06-18T11:59:36Z) - DELRec: Distilling Sequential Pattern to Enhance LLM-based Recommendation [3.5113201254928117]
Sequential recommendation (SR) tasks enhance recommendation accuracy by capturing the connection between users' past interactions and their changing preferences.
Conventional models often focus solely on capturing sequential patterns within the training data, neglecting the broader context and semantic information embedded in item titles from external sources.
DelRec aims to extract knowledge from SR models and enable LLMs to easily comprehend and utilize this supplementary information for more effective sequential recommendations.
arXiv Detail & Related papers (2024-06-17T02:47:09Z) - Integrating Large Language Models with Graphical Session-Based
Recommendation [8.086277931395212]
We introduce large language models with graphical Session-Based recommendation, named LLMGR.
This framework bridges the gap by harmoniously integrating LLMs with Graph Neural Networks (GNNs) for SBR tasks.
This integration seeks to leverage the complementary strengths of LLMs in natural language understanding and GNNs in relational data processing.
arXiv Detail & Related papers (2024-02-26T12:55:51Z) - DRDT: Dynamic Reflection with Divergent Thinking for LLM-based
Sequential Recommendation [53.62727171363384]
We introduce a novel reasoning principle: Dynamic Reflection with Divergent Thinking.
Our methodology is dynamic reflection, a process that emulates human learning through probing, critiquing, and reflecting.
We evaluate our approach on three datasets using six pre-trained LLMs.
arXiv Detail & Related papers (2023-12-18T16:41:22Z) - SR-GCL: Session-Based Recommendation with Global Context Enhanced
Augmentation in Contrastive Learning [5.346468677221906]
Session-based recommendations aim to predict the next behavior of users based on ongoing sessions.
Recent research has applied graph neural networks with an attention mechanism to capture complicated item transitions.
We propose SR-GCL, a novel contrastive learning framework for a session-based recommendation.
arXiv Detail & Related papers (2022-09-22T06:18:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.