Reason first, then respond: Modular Generation for Knowledge-infused
Dialogue
- URL: http://arxiv.org/abs/2111.05204v1
- Date: Tue, 9 Nov 2021 15:29:43 GMT
- Title: Reason first, then respond: Modular Generation for Knowledge-infused
Dialogue
- Authors: Leonard Adolphs, Kurt Shuster, Jack Urbanek, Arthur Szlam, Jason
Weston
- Abstract summary: Large language models can produce fluent dialogue but often hallucinate factual inaccuracies.
We propose a modular model, Knowledge to Response, for incorporating knowledge into conversational agents.
In detailed experiments, we find that such a model hallucinates less in knowledge-grounded dialogue tasks.
- Score: 43.64093692715295
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Large language models can produce fluent dialogue but often hallucinate
factual inaccuracies. While retrieval-augmented models help alleviate this
issue, they still face a difficult challenge of both reasoning to provide
correct knowledge and generating conversation simultaneously. In this work, we
propose a modular model, Knowledge to Response (K2R), for incorporating
knowledge into conversational agents, which breaks down this problem into two
easier steps. K2R first generates a knowledge sequence, given a dialogue
context, as an intermediate step. After this "reasoning step", the model then
attends to its own generated knowledge sequence, as well as the dialogue
context, to produce a final response. In detailed experiments, we find that
such a model hallucinates less in knowledge-grounded dialogue tasks, and has
advantages in terms of interpretability and modularity. In particular, it can
be used to fuse QA and dialogue systems together to enable dialogue agents to
give knowledgeable answers, or QA models to give conversational responses in a
zero-shot setting.
Related papers
- KPT: Keyword-guided Pre-training for Grounded Dialog Generation [82.68787152707455]
We propose KPT (guided Pre-Training), a novel self-supervised pre-training method for grounded dialog generation.
Specifically, we use a pre-trained language model to extract the most uncertain tokens in the dialog as keywords.
We conduct extensive experiments on various few-shot knowledge-grounded generation tasks, including grounding on dialog acts, knowledge graphs, persona descriptions, and Wikipedia passages.
arXiv Detail & Related papers (2022-12-04T04:05:01Z) - Dynamically Retrieving Knowledge via Query Generation for informative
dialogue response [7.196104022425989]
We design a knowledge-driven dialogue system named DRKQG (emphDynamically Retrieving Knowledge via Query Generation for informative dialogue response)
First, a time-aware mechanism is utilized to capture context information and a query can be generated for retrieving knowledge.
Then, we integrate copy Mechanism and Transformers, which allows the response generation module produces responses derived from the context and retrieved knowledge.
arXiv Detail & Related papers (2022-07-30T03:05:43Z) - Towards Large-Scale Interpretable Knowledge Graph Reasoning for Dialogue
Systems [109.16553492049441]
We propose a novel method to incorporate the knowledge reasoning capability into dialogue systems in a more scalable and generalizable manner.
To the best of our knowledge, this is the first work to have transformer models generate responses by reasoning over differentiable knowledge graphs.
arXiv Detail & Related papers (2022-03-20T17:51:49Z) - Multi-turn Dialogue Reading Comprehension with Pivot Turns and Knowledge [43.352833140317486]
Multi-turn dialogue reading comprehension aims to teach machines to read dialogue contexts and solve tasks such as response selection and answering questions.
This work makes the first attempt to tackle the above two challenges by extracting substantially important turns as pivot utterances.
We propose a pivot-oriented deep selection model (PoDS) on top of the Transformer-based language models for dialogue comprehension.
arXiv Detail & Related papers (2021-02-10T15:00:12Z) - The Adapter-Bot: All-In-One Controllable Conversational Model [66.48164003532484]
We propose a dialogue model that uses a fixed backbone model such as DialGPT and triggers on-demand dialogue skills via different adapters.
Depending on the skills, the model is able to process multiple knowledge types, such as text, tables, and emphatic responses.
We evaluate our model using automatic evaluation by comparing it with existing state-of-the-art conversational models.
arXiv Detail & Related papers (2020-08-28T10:59:31Z) - Knowledge Injection into Dialogue Generation via Language Models [85.65843021510521]
InjK is a two-stage approach to inject knowledge into a dialogue generation model.
First, we train a large-scale language model and query it as textual knowledge.
Second, we frame a dialogue generation model to sequentially generate textual knowledge and a corresponding response.
arXiv Detail & Related papers (2020-04-30T07:31:24Z) - Modality-Balanced Models for Visual Dialogue [102.35406085738325]
The Visual Dialog task requires a model to exploit both image and conversational context information to generate the next response to the dialogue.
We show that previous joint-modality (history and image) models over-rely on and are more prone to memorizing the dialogue history.
We present methods for this integration of the two models, via ensemble and consensus dropout fusion with shared parameters.
arXiv Detail & Related papers (2020-01-17T14:57:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.