A Large Language Model Enhanced Conversational Recommender System
- URL: http://arxiv.org/abs/2308.06212v1
- Date: Fri, 11 Aug 2023 16:30:44 GMT
- Title: A Large Language Model Enhanced Conversational Recommender System
- Authors: Yue Feng, Shuchang Liu, Zhenghai Xue, Qingpeng Cai, Lantao Hu, Peng
Jiang, Kun Gai, Fei Sun
- Abstract summary: Conversational recommender systems (CRSs) aim to recommend high-quality items to users through a dialogue interface.
To develop effective CRSs, there are some challenges: 1) how to properly manage sub-tasks; 2) how to effectively solve different sub-tasks; and 3) how to correctly generate responses that interact with users.
Recently, Large Language Models (LLMs) have exhibited an unprecedented ability to reason and generate, presenting a new opportunity to develop more powerful CRSs.
- Score: 25.18571087071163
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Conversational recommender systems (CRSs) aim to recommend high-quality items
to users through a dialogue interface. It usually contains multiple sub-tasks,
such as user preference elicitation, recommendation, explanation, and item
information search. To develop effective CRSs, there are some challenges: 1)
how to properly manage sub-tasks; 2) how to effectively solve different
sub-tasks; and 3) how to correctly generate responses that interact with users.
Recently, Large Language Models (LLMs) have exhibited an unprecedented ability
to reason and generate, presenting a new opportunity to develop more powerful
CRSs. In this work, we propose a new LLM-based CRS, referred to as LLMCRS, to
address the above challenges. For sub-task management, we leverage the
reasoning ability of LLM to effectively manage sub-task. For sub-task solving,
we collaborate LLM with expert models of different sub-tasks to achieve the
enhanced performance. For response generation, we utilize the generation
ability of LLM as a language interface to better interact with users.
Specifically, LLMCRS divides the workflow into four stages: sub-task detection,
model matching, sub-task execution, and response generation. LLMCRS also
designs schema-based instruction, demonstration-based instruction, dynamic
sub-task and model matching, and summary-based generation to instruct LLM to
generate desired results in the workflow. Finally, to adapt LLM to
conversational recommendations, we also propose to fine-tune LLM with
reinforcement learning from CRSs performance feedback, referred to as RLPF.
Experimental results on benchmark datasets show that LLMCRS with RLPF
outperforms the existing methods.
Related papers
- Beyond Inter-Item Relations: Dynamic Adaption for Enhancing LLM-Based Sequential Recommendation [83.87767101732351]
Sequential recommender systems (SRS) predict the next items that users may prefer based on user historical interaction sequences.
Inspired by the rise of large language models (LLMs) in various AI applications, there is a surge of work on LLM-based SRS.
We propose DARec, a sequential recommendation model built on top of coarse-grained adaption for capturing inter-item relations.
arXiv Detail & Related papers (2024-08-14T10:03:40Z) - Q*: Improving Multi-step Reasoning for LLMs with Deliberative Planning [53.6472920229013]
Large Language Models (LLMs) have demonstrated impressive capability in many natural language tasks.
LLMs are prone to produce errors, hallucinations and inconsistent statements when performing multi-step reasoning.
We introduce Q*, a framework for guiding LLMs decoding process with deliberative planning.
arXiv Detail & Related papers (2024-06-20T13:08:09Z) - A Practice-Friendly LLM-Enhanced Paradigm with Preference Parsing for Sequential Recommendation [15.153844486572932]
This paper proposes a practice-friendly LLM-enhanced paradigm with preference parsing (P2Rec) for sequential recommender systems (SRS)
Specifically, in the information reconstruction stage, we design a new user-level SFT task for collaborative information injection with the assistance of a pre-trained SRS model.
Our goal is to let LLM learn to reconstruct a corresponding prior preference distribution from each user's interaction sequence.
arXiv Detail & Related papers (2024-06-01T07:18:56Z) - Small LLMs Are Weak Tool Learners: A Multi-LLM Agent [73.54562551341454]
Large Language Model (LLM) agents significantly extend the capabilities of standalone LLMs.
We propose a novel approach that decomposes the aforementioned capabilities into a planner, caller, and summarizer.
This modular framework facilitates individual updates and the potential use of smaller LLMs for building each capability.
arXiv Detail & Related papers (2024-01-14T16:17:07Z) - Recommender AI Agent: Integrating Large Language Models for Interactive
Recommendations [53.76682562935373]
We introduce an efficient framework called textbfInteRecAgent, which employs LLMs as the brain and recommender models as tools.
InteRecAgent achieves satisfying performance as a conversational recommender system, outperforming general-purpose LLMs.
arXiv Detail & Related papers (2023-08-31T07:36:44Z) - Enabling Intelligent Interactions between an Agent and an LLM: A Reinforcement Learning Approach [31.6589518077397]
Large language models (LLMs) encode a vast amount of world knowledge acquired from massive text datasets.
LLMs can assist an embodied agent in solving complex sequential decision making tasks by providing high-level instructions.
We propose When2Ask, a reinforcement learning based approach that learns when it is necessary to query LLMs for high-level instructions.
arXiv Detail & Related papers (2023-06-06T11:49:09Z) - Low-code LLM: Graphical User Interface over Large Language Models [115.08718239772107]
This paper introduces a novel human-LLM interaction framework, Low-code LLM.
It incorporates six types of simple low-code visual programming interactions to achieve more controllable and stable responses.
We highlight three advantages of the low-code LLM: user-friendly interaction, controllable generation, and wide applicability.
arXiv Detail & Related papers (2023-04-17T09:27:40Z) - Check Your Facts and Try Again: Improving Large Language Models with
External Knowledge and Automated Feedback [127.75419038610455]
Large language models (LLMs) are able to generate human-like, fluent responses for many downstream tasks.
This paper proposes a LLM-Augmenter system, which augments a black-box LLM with a set of plug-and-play modules.
arXiv Detail & Related papers (2023-02-24T18:48:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.