Contextual Knowledge Learning For Dialogue Generation
- URL: http://arxiv.org/abs/2305.18200v1
- Date: Mon, 29 May 2023 16:54:10 GMT
- Title: Contextual Knowledge Learning For Dialogue Generation
- Authors: Wen Zheng, Natasa Milic-Frayling, Ke Zhou
- Abstract summary: We present a novel approach to context and knowledge weighting as an integral part of model training.
We guide the model training through a Contextual Knowledge Learning process which involves Latent Vectors for context and knowledge.
- Score: 13.671946960656467
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Incorporating conversational context and knowledge into dialogue generation
models has been essential for improving the quality of the generated responses.
The context, comprising utterances from previous dialogue exchanges, is used as
a source of content for response generation and as a means of selecting
external knowledge. However, to avoid introducing irrelevant content, it is key
to enable fine-grained scoring of context and knowledge. In this paper, we
present a novel approach to context and knowledge weighting as an integral part
of model training. We guide the model training through a Contextual Knowledge
Learning (CKL) process which involves Latent Vectors for context and knowledge,
respectively. CKL Latent Vectors capture the relationship between context,
knowledge, and responses through weak supervision and enable differential
weighting of context utterances and knowledge sentences during the training
process. Experiments with two standard datasets and human evaluation
demonstrate that CKL leads to a significant improvement compared with the
performance of six strong baseline models and shows robustness with regard to
reduced sizes of training sets.
Related papers
- Large Language Models are Limited in Out-of-Context Knowledge Reasoning [65.72847298578071]
Large Language Models (LLMs) possess extensive knowledge and strong capabilities in performing in-context reasoning.
This paper focuses on a significant aspect of out-of-context reasoning: Out-of-Context Knowledge Reasoning (OCKR), which is to combine multiple knowledge to infer new knowledge.
arXiv Detail & Related papers (2024-06-11T15:58:59Z) - Knowledge Graphs and Pre-trained Language Models enhanced Representation Learning for Conversational Recommender Systems [58.561904356651276]
We introduce the Knowledge-Enhanced Entity Representation Learning (KERL) framework to improve the semantic understanding of entities for Conversational recommender systems.
KERL uses a knowledge graph and a pre-trained language model to improve the semantic understanding of entities.
KERL achieves state-of-the-art results in both recommendation and response generation tasks.
arXiv Detail & Related papers (2023-12-18T06:41:23Z) - RT-KGD: Relation Transition Aware Knowledge-Grounded Dialogue Generation [20.37399983466163]
We propose a Relation Transition aware Knowledge-Grounded Dialogue Generation model (RT-KGD)
Specifically, inspired by the latent logic of human conversation, our model integrates dialogue-level relation transition regularities with turn-level entity semantic information.
In this manner, the interaction between knowledge is considered to produce abundant clues for predicting the appropriate knowledge and generating coherent responses.
arXiv Detail & Related papers (2022-07-17T16:07:38Z) - Multimodal Dialog Systems with Dual Knowledge-enhanced Generative Pretrained Language Model [63.461030694700014]
We propose a novel dual knowledge-enhanced generative pretrained language model for multimodal task-oriented dialog systems (DKMD)
The proposed DKMD consists of three key components: dual knowledge selection, dual knowledge-enhanced context learning, and knowledge-enhanced response generation.
Experiments on a public dataset verify the superiority of the proposed DKMD over state-of-the-art competitors.
arXiv Detail & Related papers (2022-07-16T13:02:54Z) - Learning to Express in Knowledge-Grounded Conversation [62.338124154016825]
We consider two aspects of knowledge expression, namely the structure of the response and style of the content in each part.
We propose a segmentation-based generation model and optimize the model by a variational approach to discover the underlying pattern of knowledge expression in a response.
arXiv Detail & Related papers (2022-04-12T13:43:47Z) - A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded
Dialogue Generation [0.9926500244448218]
We propose a novel three-stage learning framework based on weakly supervised learning which benefits from large scale ungrounded dialogues and unstructured knowledge base.
Our approach can outperform other state-of-the-art methods with less training data, and even in zero-resource scenario, our approach still performs well.
arXiv Detail & Related papers (2021-09-09T08:32:02Z) - Contextualized Knowledge-aware Attentive Neural Network: Enhancing
Answer Selection with Knowledge [77.77684299758494]
We extensively investigate approaches to enhancing the answer selection model with external knowledge from knowledge graph (KG)
First, we present a context-knowledge interaction learning framework, Knowledge-aware Neural Network (KNN), which learns the QA sentence representations by considering a tight interaction with the external knowledge from KG and the textual information.
To handle the diversity and complexity of KG information, we propose a Contextualized Knowledge-aware Attentive Neural Network (CKANN), which improves the knowledge representation learning with structure information via a customized Graph Convolutional Network (GCN) and comprehensively learns context-based and knowledge-based sentence representation via
arXiv Detail & Related papers (2021-04-12T05:52:20Z) - Zero-Resource Knowledge-Grounded Dialogue Generation [29.357221039484568]
We propose representing the knowledge that bridges a context and a response and the way that the knowledge is expressed as latent variables.
We show that our model can achieve comparable performance with state-of-the-art methods that rely on knowledge-grounded dialogues for training.
arXiv Detail & Related papers (2020-08-29T05:48:32Z) - Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue [51.513276162736844]
We propose a sequential latent variable model as the first approach to this matter.
The model named sequential knowledge transformer (SKT) can keep track of the prior and posterior distribution over knowledge.
arXiv Detail & Related papers (2020-02-18T11:59:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.