A Large Language Model Enhanced Conversational Recommender System
- URL: http://arxiv.org/abs/2308.06212v1
- Date: Fri, 11 Aug 2023 16:30:44 GMT
- Title: A Large Language Model Enhanced Conversational Recommender System
- Authors: Yue Feng, Shuchang Liu, Zhenghai Xue, Qingpeng Cai, Lantao Hu, Peng
Jiang, Kun Gai, Fei Sun
- Abstract summary: Conversational recommender systems (CRSs) aim to recommend high-quality items to users through a dialogue interface.
To develop effective CRSs, there are some challenges: 1) how to properly manage sub-tasks; 2) how to effectively solve different sub-tasks; and 3) how to correctly generate responses that interact with users.
Recently, Large Language Models (LLMs) have exhibited an unprecedented ability to reason and generate, presenting a new opportunity to develop more powerful CRSs.
- Score: 25.18571087071163
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Conversational recommender systems (CRSs) aim to recommend high-quality items
to users through a dialogue interface. It usually contains multiple sub-tasks,
such as user preference elicitation, recommendation, explanation, and item
information search. To develop effective CRSs, there are some challenges: 1)
how to properly manage sub-tasks; 2) how to effectively solve different
sub-tasks; and 3) how to correctly generate responses that interact with users.
Recently, Large Language Models (LLMs) have exhibited an unprecedented ability
to reason and generate, presenting a new opportunity to develop more powerful
CRSs. In this work, we propose a new LLM-based CRS, referred to as LLMCRS, to
address the above challenges. For sub-task management, we leverage the
reasoning ability of LLM to effectively manage sub-task. For sub-task solving,
we collaborate LLM with expert models of different sub-tasks to achieve the
enhanced performance. For response generation, we utilize the generation
ability of LLM as a language interface to better interact with users.
Specifically, LLMCRS divides the workflow into four stages: sub-task detection,
model matching, sub-task execution, and response generation. LLMCRS also
designs schema-based instruction, demonstration-based instruction, dynamic
sub-task and model matching, and summary-based generation to instruct LLM to
generate desired results in the workflow. Finally, to adapt LLM to
conversational recommendations, we also propose to fine-tune LLM with
reinforcement learning from CRSs performance feedback, referred to as RLPF.
Experimental results on benchmark datasets show that LLMCRS with RLPF
outperforms the existing methods.
Related papers
- Q*: Improving Multi-step Reasoning for LLMs with Deliberative Planning [53.6472920229013]
Large Language Models (LLMs) have demonstrated impressive capability in many natural language tasks.
LLMs are prone to produce errors, hallucinations and inconsistent statements when performing multi-step reasoning.
We introduce Q*, a framework for guiding LLMs decoding process with deliberative planning.
arXiv Detail & Related papers (2024-06-20T13:08:09Z) - A Multi-Agent Conversational Recommender System [47.49330334611104]
We propose Multi-Agent Conversational Recommender System (MACRS)
First, we design a multi-agent act planning framework, which can control the dialogue flow based on four LLM-based agents.
Second, we propose a user feedback-aware reflection mechanism which leverages user feedback to reason errors made in previous turns to adjust the dialogue act planning.
arXiv Detail & Related papers (2024-02-02T04:20:13Z) - MLLMReID: Multimodal Large Language Model-based Person Re-identification [14.68436005777866]
Multimodal large language models (MLLM) have achieved satisfactory results in many tasks.
This paper will investigate how to adapt them for the task of ReID.
arXiv Detail & Related papers (2024-01-24T03:07:26Z) - Small LLMs Are Weak Tool Learners: A Multi-LLM Agent [73.54562551341454]
Large Language Model (LLM) agents significantly extend the capabilities of standalone LLMs.
We propose a novel approach that decomposes the aforementioned capabilities into a planner, caller, and summarizer.
This modular framework facilitates individual updates and the potential use of smaller LLMs for building each capability.
arXiv Detail & Related papers (2024-01-14T16:17:07Z) - Rephrase and Respond: Let Large Language Models Ask Better Questions for Themselves [57.974103113675795]
We present a method named Rephrase and Respond' (RaR) which allows Large Language Models to rephrase and expand questions posed by humans.
RaR serves as a simple yet effective prompting method for improving performance.
We show that RaR is complementary to the popular Chain-of-Thought (CoT) methods, both theoretically and empirically.
arXiv Detail & Related papers (2023-11-07T18:43:34Z) - Recommender AI Agent: Integrating Large Language Models for Interactive
Recommendations [53.76682562935373]
We introduce an efficient framework called textbfInteRecAgent, which employs LLMs as the brain and recommender models as tools.
InteRecAgent achieves satisfying performance as a conversational recommender system, outperforming general-purpose LLMs.
arXiv Detail & Related papers (2023-08-31T07:36:44Z) - Enabling Intelligent Interactions between an Agent and an LLM: A Reinforcement Learning Approach [31.6589518077397]
Large language models (LLMs) encode a vast amount of world knowledge acquired from massive text datasets.
LLMs can assist an embodied agent in solving complex sequential decision making tasks by providing high-level instructions.
We propose When2Ask, a reinforcement learning based approach that learns when it is necessary to query LLMs for high-level instructions.
arXiv Detail & Related papers (2023-06-06T11:49:09Z) - Low-code LLM: Graphical User Interface over Large Language Models [115.08718239772107]
This paper introduces a novel human-LLM interaction framework, Low-code LLM.
It incorporates six types of simple low-code visual programming interactions to achieve more controllable and stable responses.
We highlight three advantages of the low-code LLM: user-friendly interaction, controllable generation, and wide applicability.
arXiv Detail & Related papers (2023-04-17T09:27:40Z) - Check Your Facts and Try Again: Improving Large Language Models with
External Knowledge and Automated Feedback [127.75419038610455]
Large language models (LLMs) are able to generate human-like, fluent responses for many downstream tasks.
This paper proposes a LLM-Augmenter system, which augments a black-box LLM with a set of plug-and-play modules.
arXiv Detail & Related papers (2023-02-24T18:48:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.