Enhanced Knowledge Selection for Grounded Dialogues via Document
Semantic Graphs
- URL: http://arxiv.org/abs/2206.07296v1
- Date: Wed, 15 Jun 2022 04:51:32 GMT
- Title: Enhanced Knowledge Selection for Grounded Dialogues via Document
Semantic Graphs
- Authors: Sha Li, Madhi Namazifar, Di Jin, Mohit Bansal, Heng Ji, Yang Liu,
Dilek Hakkani-Tur
- Abstract summary: We propose to automatically convert background knowledge documents into document semantic graphs.
Our document semantic graphs preserve sentence-level information through the use of sentence nodes and provide concept connections between sentences.
Our experiments show that our semantic graph-based knowledge selection improves over sentence selection baselines for both the knowledge selection task and the end-to-end response generation task on HollE.
- Score: 123.50636090341236
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Providing conversation models with background knowledge has been shown to
make open-domain dialogues more informative and engaging. Existing models treat
knowledge selection as a sentence ranking or classification problem where each
sentence is handled individually, ignoring the internal semantic connection
among sentences in the background document. In this work, we propose to
automatically convert the background knowledge documents into document semantic
graphs and then perform knowledge selection over such graphs. Our document
semantic graphs preserve sentence-level information through the use of sentence
nodes and provide concept connections between sentences. We jointly apply
multi-task learning for sentence-level and concept-level knowledge selection
and show that it improves sentence-level selection. Our experiments show that
our semantic graph-based knowledge selection improves over sentence selection
baselines for both the knowledge selection task and the end-to-end response
generation task on HollE and improves generalization on unseen topics in WoW.
Related papers
- Word and Phrase Features in Graph Convolutional Network for Automatic Question Classification [0.7405975743268344]
We propose a novel approach leveraging graph convolutional networks (GCNs) to better model the inherent structure of questions.
By representing questions as graphs, our method allows GCNs to learn from the interconnected nature of language more effectively.
Our findings demonstrate that GCNs, augmented with phrase-based features, offer a promising solution for more accurate and context-aware question classification.
arXiv Detail & Related papers (2024-09-04T07:13:30Z) - Knowledge Graphs and Pre-trained Language Models enhanced Representation Learning for Conversational Recommender Systems [58.561904356651276]
We introduce the Knowledge-Enhanced Entity Representation Learning (KERL) framework to improve the semantic understanding of entities for Conversational recommender systems.
KERL uses a knowledge graph and a pre-trained language model to improve the semantic understanding of entities.
KERL achieves state-of-the-art results in both recommendation and response generation tasks.
arXiv Detail & Related papers (2023-12-18T06:41:23Z) - Conversational Semantic Parsing using Dynamic Context Graphs [68.72121830563906]
We consider the task of conversational semantic parsing over general purpose knowledge graphs (KGs) with millions of entities, and thousands of relation-types.
We focus on models which are capable of interactively mapping user utterances into executable logical forms.
arXiv Detail & Related papers (2023-05-04T16:04:41Z) - The Short Text Matching Model Enhanced with Knowledge via Contrastive
Learning [8.350445155753167]
This paper proposes a short Text Matching model that combines contrastive learning and external knowledge.
To avoid noise, we use keywords as the main semantics of the original sentence to retrieve corresponding knowledge words in the knowledge base.
Our designed model achieves state-of-the-art performance on two publicly available Chinese Text Matching datasets.
arXiv Detail & Related papers (2023-04-08T03:24:05Z) - Multi-Sentence Knowledge Selection in Open-Domain Dialogue [11.936691632841388]
We evaluate the existing state of open-domain conversation knowledge selection.
We create an augmented dataset based on the Wizard of Wikipedia (WOW) corpus.
WOW++ averages 8 relevant knowledge sentences per dialogue context.
arXiv Detail & Related papers (2022-03-01T22:07:05Z) - Variational Learning for Unsupervised Knowledge Grounded Dialogs [6.761874595503588]
Recent methods for knowledge grounded dialogs generate responses by incorporating information from an external textual document.
We develop a variational approach to the above technique wherein, we instead maximize the Evidence Lower bound (ELBO)
To the best of our knowledge we are the first to apply variational training for open-scale unsupervised knowledge grounded dialog systems.
arXiv Detail & Related papers (2021-11-23T13:41:03Z) - Reasoning in Dialog: Improving Response Generation by Context Reading
Comprehension [49.92173751203827]
In multi-turn dialog, utterances do not always take the full form of sentences.
We propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question.
arXiv Detail & Related papers (2020-12-14T10:58:01Z) - Exploiting Structured Knowledge in Text via Graph-Guided Representation
Learning [73.0598186896953]
We present two self-supervised tasks learning over raw text with the guidance from knowledge graphs.
Building upon entity-level masked language models, our first contribution is an entity masking scheme.
In contrast to existing paradigms, our approach uses knowledge graphs implicitly, only during pre-training.
arXiv Detail & Related papers (2020-04-29T14:22:42Z) - Knowledge-graph based Proactive Dialogue Generation with Improved
Meta-Learning [0.0]
We propose a knowledge graph based proactive dialogue generation model (KgDg) with three components.
For knowledge triplets embedding and selection, we formulate it as a problem of sentence embedding to better capture semantic information.
Our improved MAML algorithm is capable of learning general features from a limited number of knowledge graphs.
arXiv Detail & Related papers (2020-04-19T08:41:12Z) - Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue [51.513276162736844]
We propose a sequential latent variable model as the first approach to this matter.
The model named sequential knowledge transformer (SKT) can keep track of the prior and posterior distribution over knowledge.
arXiv Detail & Related papers (2020-02-18T11:59:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.