Finetuning Large-Scale Pre-trained Language Models for Conversational
Recommendation with Knowledge Graph
- URL: http://arxiv.org/abs/2110.07477v1
- Date: Thu, 14 Oct 2021 15:49:48 GMT
- Title: Finetuning Large-Scale Pre-trained Language Models for Conversational
Recommendation with Knowledge Graph
- Authors: Lingzhi Wang, Huang Hu, Lei Sha, Can Xu, Kam-Fai Wong, Daxin Jiang
- Abstract summary: We present a pre-trained language model (PLM) based framework called RID conversational recommender system (CRS)
RID significantly outperforms the state-of-the-art methods on both evaluations of dialogue and recommendation.
- Score: 35.033130888779226
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present a pre-trained language model (PLM) based framework
called RID for conversational recommender system (CRS). RID finetunes the
large-scale PLMs such as DialoGPT, together with a pre-trained Relational Graph
Convolutional Network (RGCN) to encode the node representations of an
item-oriented knowledge graph. The former aims to generate fluent and diverse
dialogue responses based on the strong language generation ability of PLMs,
while the latter is to facilitate the item recommendation by learning better
node embeddings on the structural knowledge base. To unify two modules of
dialogue generation and item recommendation into a PLMs-based framework, we
expand the generation vocabulary of PLMs to include an extra item vocabulary,
and introduces a vocabulary pointer to control when to recommend target items
in the generation process. Extensive experiments on the benchmark dataset
ReDial show RID significantly outperforms the state-of-the-art methods on both
evaluations of dialogue and recommendation.
Related papers
- Align-SLM: Textless Spoken Language Models with Reinforcement Learning from AI Feedback [50.84142264245052]
This work introduces the Align-SLM framework to enhance the semantic understanding of textless Spoken Language Models (SLMs)
Our approach generates multiple speech continuations from a given prompt and uses semantic metrics to create preference data for Direct Preference Optimization (DPO)
We evaluate the framework using ZeroSpeech 2021 benchmarks for lexical and syntactic modeling, the spoken version of the StoryCloze dataset for semantic coherence, and other speech generation metrics, including the GPT4-o score and human evaluation.
arXiv Detail & Related papers (2024-11-04T06:07:53Z) - Enhancing Collaborative Semantics of Language Model-Driven Recommendations via Graph-Aware Learning [10.907949155931474]
Large Language Models (LLMs) are increasingly prominent in the recommendation systems domain.
Gal-Rec enhances the understanding of user-item collaborative semantics by imitating the intent of Graph Neural Networks (GNNs)
Gal-Rec significantly enhances the comprehension of collaborative semantics, and improves recommendation performance.
arXiv Detail & Related papers (2024-06-19T05:50:15Z) - Parameter-Efficient Conversational Recommender System as a Language
Processing Task [52.47087212618396]
Conversational recommender systems (CRS) aim to recommend relevant items to users by eliciting user preference through natural language conversation.
Prior work often utilizes external knowledge graphs for items' semantic information, a language model for dialogue generation, and a recommendation module for ranking relevant items.
In this paper, we represent items in natural language and formulate CRS as a natural language processing task.
arXiv Detail & Related papers (2024-01-25T14:07:34Z) - Knowledge Graphs and Pre-trained Language Models enhanced Representation Learning for Conversational Recommender Systems [58.561904356651276]
We introduce the Knowledge-Enhanced Entity Representation Learning (KERL) framework to improve the semantic understanding of entities for Conversational recommender systems.
KERL uses a knowledge graph and a pre-trained language model to improve the semantic understanding of entities.
KERL achieves state-of-the-art results in both recommendation and response generation tasks.
arXiv Detail & Related papers (2023-12-18T06:41:23Z) - Rethinking the Evaluation for Conversational Recommendation in the Era
of Large Language Models [115.7508325840751]
The recent success of large language models (LLMs) has shown great potential to develop more powerful conversational recommender systems (CRSs)
In this paper, we embark on an investigation into the utilization of ChatGPT for conversational recommendation, revealing the inadequacy of the existing evaluation protocol.
We propose an interactive Evaluation approach based on LLMs named iEvaLM that harnesses LLM-based user simulators.
arXiv Detail & Related papers (2023-05-22T15:12:43Z) - Towards Unified Conversational Recommender Systems via
Knowledge-Enhanced Prompt Learning [89.64215566478931]
Conversational recommender systems (CRS) aim to proactively elicit user preference and recommend high-quality items through natural language conversations.
To develop an effective CRS, it is essential to seamlessly integrate the two modules.
We propose a unified CRS model named UniCRS based on knowledge-enhanced prompt learning.
arXiv Detail & Related papers (2022-06-19T09:21:27Z) - Improving Conversational Recommendation Systems' Quality with
Context-Aware Item Meta Information [42.88448098873448]
Conversational recommendation systems (CRS) engage with users by inferring user preferences from dialog history.
Previous CRSs use knowledge graph (KG) based recommendation modules and integrate KG with language models for response generation.
We propose a simple yet effective architecture comprising a pre-trained language model (PLM) and an item metadata encoder.
arXiv Detail & Related papers (2021-12-15T14:12:48Z) - DialogBERT: Discourse-Aware Response Generation via Learning to Recover
and Rank Utterances [18.199473005335093]
This paper presents DialogBERT, a novel conversational response generation model that enhances previous PLM-based dialogue models.
To efficiently capture the discourse-level coherence among utterances, we propose two training objectives, including masked utterance regression.
Experiments on three multi-turn conversation datasets show that our approach remarkably outperforms the baselines.
arXiv Detail & Related papers (2020-12-03T09:06:23Z) - Improving Conversational Recommender Systems via Knowledge Graph based
Semantic Fusion [77.21442487537139]
Conversational recommender systems (CRS) aim to recommend high-quality items to users through interactive conversations.
First, the conversation data itself lacks of sufficient contextual information for accurately understanding users' preference.
Second, there is a semantic gap between natural language expression and item-level user preference.
arXiv Detail & Related papers (2020-07-08T11:14:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.