LLM4SBR: A Lightweight and Effective Framework for Integrating Large
Language Models in Session-based Recommendation
- URL: http://arxiv.org/abs/2402.13840v1
- Date: Wed, 21 Feb 2024 14:38:02 GMT
- Title: LLM4SBR: A Lightweight and Effective Framework for Integrating Large
Language Models in Session-based Recommendation
- Authors: Shutong Qiao, Chen Gao, Junhao Wen, Wei Zhou, Qun Luo, Peixuan Chen
and Yong Li
- Abstract summary: Traditional session-based recommendation (SBR) utilizes session behavior sequences from anonymous users for recommendation.
We propose the LLM Integration Framework for SBR (LLM4SBR) as a lightweight and plug-and-play framework.
We conducted experiments on two real-world datasets, and the results demonstrate that LLM4SBR significantly improves the performance of traditional SBR models.
- Score: 27.922143384779563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traditional session-based recommendation (SBR) utilizes session behavior
sequences from anonymous users for recommendation. Although this strategy is
highly efficient, it sacrifices the inherent semantic information of the items,
making it difficult for the model to understand the true intent of the session
and resulting in a lack of interpretability in the recommended results.
Recently, large language models (LLMs) have flourished across various domains,
offering a glimpse of hope in addressing the aforementioned challenges.
Inspired by the impact of LLMs, research exploring the integration of LLMs with
the Recommender system (RS) has surged like mushrooms after rain. However,
constrained by high time and space costs, as well as the brief and anonymous
nature of session data, the first LLM recommendation framework suitable for
industrial deployment has yet to emerge in the field of SBR. To address the
aforementioned challenges, we have proposed the LLM Integration Framework for
SBR (LLM4SBR). Serving as a lightweight and plug-and-play framework, LLM4SBR
adopts a two-step strategy. Firstly, we transform session data into a bimodal
form of text and behavior. In the first step, leveraging the inferential
capabilities of LLMs, we conduct inference on session text data from different
perspectives and design the component for auxiliary enhancement. In the second
step, the SBR model is trained on behavior data, aligning and averaging two
modal session representations from different perspectives. Finally, we fuse
session representations from different perspectives and modalities as the
ultimate session representation for recommendation. We conducted experiments on
two real-world datasets, and the results demonstrate that LLM4SBR significantly
improves the performance of traditional SBR models and is highly lightweight
and efficient, making it suitable for industrial deployment.
Related papers
- Understand What LLM Needs: Dual Preference Alignment for Retrieval-Augmented Generation [64.7982176398485]
Retrieval-augmented generation (RAG) has demonstrated effectiveness in mitigating the hallucination problem of large language models (LLMs)
We propose DPA-RAG, a universal framework designed to align diverse knowledge preferences within RAG systems.
arXiv Detail & Related papers (2024-06-26T18:26:53Z) - LLM4MSR: An LLM-Enhanced Paradigm for Multi-Scenario Recommendation [45.31960122494715]
We propose an efficient interpretable large language model (LLM)-enhanced paradigm LLM4MSR.
Specifically, we first leverage LLM to uncover multi-level knowledge including scenario correlations and users' cross-scenario interests.
Our experiments on KuaiSAR-small, KuaiSAR, and Amazon datasets validate two significant advantages of LLM4MSR.
arXiv Detail & Related papers (2024-06-18T11:59:36Z) - DELRec: Distilling Sequential Pattern to Enhance LLM-based Recommendation [3.5113201254928117]
Sequential recommendation (SR) tasks enhance recommendation accuracy by capturing the connection between users' past interactions and their changing preferences.
Conventional models often focus solely on capturing sequential patterns within the training data, neglecting the broader context and semantic information embedded in item titles from external sources.
DelRec aims to extract knowledge from SR models and enable LLMs to easily comprehend and utilize this supplementary information for more effective sequential recommendations.
arXiv Detail & Related papers (2024-06-17T02:47:09Z) - Mitigating Catastrophic Forgetting in Large Language Models with Self-Synthesized Rehearsal [49.24054920683246]
Large language models (LLMs) suffer from catastrophic forgetting during continual learning.
We propose a framework called Self-Synthesized Rehearsal (SSR) that uses the LLM to generate synthetic instances for rehearsal.
arXiv Detail & Related papers (2024-03-02T16:11:23Z) - Integrating Large Language Models with Graphical Session-Based
Recommendation [8.086277931395212]
We introduce large language models with graphical Session-Based recommendation, named LLMGR.
This framework bridges the gap by harmoniously integrating LLMs with Graph Neural Networks (GNNs) for SBR tasks.
This integration seeks to leverage the complementary strengths of LLMs in natural language understanding and GNNs in relational data processing.
arXiv Detail & Related papers (2024-02-26T12:55:51Z) - An Embarrassingly Simple Approach for LLM with Strong ASR Capacity [56.30595787061546]
We focus on solving one of the most important tasks in the field of speech processing, with speech foundation encoders and large language models (LLM)
Recent works have complex designs such as compressing the output temporally for the speech encoder, tackling modal alignment for the projector, and utilizing parameter-efficient fine-tuning for the LLM.
We found that delicate designs are not necessary, while an embarrassingly simple composition of off-the-shelf speech encoder, LLM, and the only trainable linear projector is competent for the ASR task.
arXiv Detail & Related papers (2024-02-13T23:25:04Z) - DRDT: Dynamic Reflection with Divergent Thinking for LLM-based
Sequential Recommendation [53.62727171363384]
We introduce a novel reasoning principle: Dynamic Reflection with Divergent Thinking.
Our methodology is dynamic reflection, a process that emulates human learning through probing, critiquing, and reflecting.
We evaluate our approach on three datasets using six pre-trained LLMs.
arXiv Detail & Related papers (2023-12-18T16:41:22Z) - Large Language Models for Intent-Driven Session Recommendations [34.64421003286209]
We introduce a novel ISR approach, utilizing the advanced reasoning capabilities of large language models (LLMs)
We introduce an innovative prompt optimization mechanism that iteratively self-reflects and adjusts prompts.
This new paradigm empowers LLMs to discern diverse user intents at a semantic level, leading to more accurate and interpretable session recommendations.
arXiv Detail & Related papers (2023-12-07T02:25:14Z) - Query-Dependent Prompt Evaluation and Optimization with Offline Inverse
RL [62.824464372594576]
We aim to enhance arithmetic reasoning ability of Large Language Models (LLMs) through zero-shot prompt optimization.
We identify a previously overlooked objective of query dependency in such optimization.
We introduce Prompt-OIRL, which harnesses offline inverse reinforcement learning to draw insights from offline prompting demonstration data.
arXiv Detail & Related papers (2023-09-13T01:12:52Z) - SR-GCL: Session-Based Recommendation with Global Context Enhanced
Augmentation in Contrastive Learning [5.346468677221906]
Session-based recommendations aim to predict the next behavior of users based on ongoing sessions.
Recent research has applied graph neural networks with an attention mechanism to capture complicated item transitions.
We propose SR-GCL, a novel contrastive learning framework for a session-based recommendation.
arXiv Detail & Related papers (2022-09-22T06:18:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.