Rome was built in 1776: A Case Study on Factual Correctness in
Knowledge-Grounded Response Generation
- URL: http://arxiv.org/abs/2110.05456v1
- Date: Mon, 11 Oct 2021 17:48:11 GMT
- Title: Rome was built in 1776: A Case Study on Factual Correctness in
Knowledge-Grounded Response Generation
- Authors: Sashank Santhanam, Behnam Hedayatnia, Spandana Gella, Aishwarya
Padmakumar, Seokhwan Kim, Yang Liu, Dilek Hakkani-Tur
- Abstract summary: We present a human annotation setup to identify three different response types.
We automatically create a new corpus called Conv-FEVER that is adapted from the Wizard of Wikipedia dataset.
- Score: 18.63673852470077
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently neural response generation models have leveraged large pre-trained
transformer models and knowledge snippets to generate relevant and informative
responses. However, this does not guarantee that generated responses are
factually correct. In this paper, we examine factual correctness in
knowledge-grounded neural response generation models. We present a human
annotation setup to identify three different response types: responses that are
factually consistent with respect to the input knowledge, responses that
contain hallucinated knowledge, and non-verifiable chitchat style responses. We
use this setup to annotate responses generated using different stateof-the-art
models, knowledge snippets, and decoding strategies. In addition, to facilitate
the development of a factual consistency detector, we automatically create a
new corpus called Conv-FEVER that is adapted from the Wizard of Wikipedia
dataset and includes factually consistent and inconsistent responses. We
demonstrate the benefit of our Conv-FEVER dataset by showing that the models
trained on this data perform reasonably well to detect factually inconsistent
responses with respect to the provided knowledge through evaluation on our
human annotated data. We will release the Conv-FEVER dataset and the human
annotated responses.
Related papers
- Contrastive Learning to Improve Retrieval for Real-world Fact Checking [84.57583869042791]
We present Contrastive Fact-Checking Reranker (CFR), an improved retriever for fact-checking complex claims.
We leverage the AVeriTeC dataset, which annotates subquestions for claims with human written answers from evidence documents.
We find a 6% improvement in veracity classification accuracy on the dataset.
arXiv Detail & Related papers (2024-10-07T00:09:50Z) - Evidence-Driven Retrieval Augmented Response Generation for Online Misinformation [18.18205773056388]
We propose retrieval augmented response generation for online misinformation (RARG)
RARG collects supporting evidence from scientific sources and generates counter-misinformation responses based on the evidences.
We propose a reward function to maximize the utilization of the retrieved evidence while maintaining the quality of the generated text.
arXiv Detail & Related papers (2024-03-22T05:05:45Z) - R-Tuning: Instructing Large Language Models to Say `I Don't Know' [66.11375475253007]
Large language models (LLMs) have revolutionized numerous domains with their impressive performance but still face their challenges.
Previous instruction tuning methods force the model to complete a sentence no matter whether the model knows the knowledge or not.
We present a new approach called Refusal-Aware Instruction Tuning (R-Tuning)
Experimental results demonstrate R-Tuning effectively improves a model's ability to answer known questions and refrain from answering unknown questions.
arXiv Detail & Related papers (2023-11-16T08:45:44Z) - Promoting Open-domain Dialogue Generation through Learning Pattern
Information between Contexts and Responses [5.936682548344234]
This paper improves the quality of generated responses by learning the implicit pattern information between contexts and responses in the training samples.
We also design a response-aware mechanism for mining the implicit pattern information between contexts and responses so that the generated replies are more diverse and approximate to human replies.
arXiv Detail & Related papers (2023-09-06T08:11:39Z) - RECKONING: Reasoning through Dynamic Knowledge Encoding [51.076603338764706]
We show that language models can answer questions by reasoning over knowledge provided as part of the context.
In these situations, the model fails to distinguish the knowledge that is necessary to answer the question.
We propose teaching the model to reason more robustly by folding the provided contextual knowledge into the model's parameters.
arXiv Detail & Related papers (2023-05-10T17:54:51Z) - Reinforcement Learning-based Counter-Misinformation Response Generation:
A Case Study of COVID-19 Vaccine Misinformation [19.245814221211415]
Non-expert ordinary users act as eyes-on-the-ground who proactively counter misinformation.
In this work, we create two novel datasets of misinformation and counter-misinformation response pairs.
We propose MisinfoCorrect, a reinforcement learning-based framework that learns to generate counter-misinformation responses.
arXiv Detail & Related papers (2023-03-11T15:55:01Z) - Visual Question Answering with Prior Class Semantics [50.845003775809836]
We show how to exploit additional information pertaining to the semantics of candidate answers.
We extend the answer prediction process with a regression objective in a semantic space.
Our method brings improvements in consistency and accuracy over a range of question types.
arXiv Detail & Related papers (2020-05-04T02:46:31Z) - A Controllable Model of Grounded Response Generation [122.7121624884747]
Current end-to-end neural conversation models inherently lack the flexibility to impose semantic control in the response generation process.
We propose a framework that we call controllable grounded response generation (CGRG)
We show that using this framework, a transformer based model with a novel inductive attention mechanism, trained on a conversation-like Reddit dataset, outperforms strong generation baselines.
arXiv Detail & Related papers (2020-05-01T21:22:08Z) - Improving Factual Consistency Between a Response and Persona Facts [64.30785349238619]
Neural models for response generation produce responses that are semantically plausible but not necessarily factually consistent with facts describing the speaker's persona.
We propose to fine-tune these models by reinforcement learning and an efficient reward function that explicitly captures the consistency between a response and persona facts as well as semantic plausibility.
arXiv Detail & Related papers (2020-04-30T18:08:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.