QA2Explanation: Generating and Evaluating Explanations for Question
Answering Systems over Knowledge Graph
- URL: http://arxiv.org/abs/2010.08323v1
- Date: Fri, 16 Oct 2020 11:32:12 GMT
- Title: QA2Explanation: Generating and Evaluating Explanations for Question
Answering Systems over Knowledge Graph
- Authors: Saeedeh Shekarpour, Abhishek Nadgeri and Kuldeep Singh
- Abstract summary: We develop an automatic approach for generating explanations during various stages of a pipeline-based QA system.
Our approach is a supervised and automatic approach which considers three classes (i.e., success, no answer, and wrong answer) for annotating the output of involved QA components.
- Score: 4.651476054353298
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the era of Big Knowledge Graphs, Question Answering (QA) systems have
reached a milestone in their performance and feasibility. However, their
applicability, particularly in specific domains such as the biomedical domain,
has not gained wide acceptance due to their "black box" nature, which hinders
transparency, fairness, and accountability of QA systems. Therefore, users are
unable to understand how and why particular questions have been answered,
whereas some others fail. To address this challenge, in this paper, we develop
an automatic approach for generating explanations during various stages of a
pipeline-based QA system. Our approach is a supervised and automatic approach
which considers three classes (i.e., success, no answer, and wrong answer) for
annotating the output of involved QA components. Upon our prediction, a
template explanation is chosen and integrated into the output of the
corresponding component. To measure the effectiveness of the approach, we
conducted a user survey as to how non-expert users perceive our generated
explanations. The results of our study show a significant increase in the four
dimensions of the human factor from the Human-computer interaction community.
Related papers
- PIE-QG: Paraphrased Information Extraction for Unsupervised Question
Generation from Small Corpora [4.721845865189576]
PIE-QG uses Open Information Extraction (OpenIE) to generate synthetic training questions from paraphrased passages.
Triples in the form of subject, predicate, object> are extracted from each passage, and questions are formed with subjects (or objects) and predicates while objects (or subjects) are considered as answers.
arXiv Detail & Related papers (2023-01-03T12:20:51Z) - Utilizing Background Knowledge for Robust Reasoning over Traffic
Situations [63.45021731775964]
We focus on a complementary research aspect of Intelligent Transportation: traffic understanding.
We scope our study to text-based methods and datasets given the abundant commonsense knowledge.
We adopt three knowledge-driven approaches for zero-shot QA over traffic situations.
arXiv Detail & Related papers (2022-12-04T09:17:24Z) - Evaluation of Question Answering Systems: Complexity of judging a
natural language [3.4771957347698583]
Question answering (QA) systems are among the most important and rapidly developing research topics in natural language processing (NLP)
This survey attempts to provide a systematic overview of the general framework of QA, QA paradigms, benchmark datasets, and assessment techniques for a quantitative evaluation of QA systems.
arXiv Detail & Related papers (2022-09-10T12:29:04Z) - Calibrating Trust of Multi-Hop Question Answering Systems with
Decompositional Probes [14.302797773412543]
Multi-hop Question Answering (QA) is a challenging task since it requires an accurate aggregation of information from multiple context paragraphs.
Recent work in multi-hop QA has shown that performance can be boosted by first decomposing the questions into simpler, single-hop questions.
We show that decomposition is an effective form of probing QA systems as well as a promising approach to explanation generation.
arXiv Detail & Related papers (2022-04-16T01:03:36Z) - Improving the Question Answering Quality using Answer Candidate
Filtering based on Natural-Language Features [117.44028458220427]
We address the problem of how the Question Answering (QA) quality of a given system can be improved.
Our main contribution is an approach capable of identifying wrong answers provided by a QA system.
In particular, our approach has shown its potential while removing in many cases the majority of incorrect answers.
arXiv Detail & Related papers (2021-12-10T11:09:44Z) - Improving Unsupervised Question Answering via Summarization-Informed
Question Generation [47.96911338198302]
Question Generation (QG) is the task of generating a plausible question for a passage, answer> pair.
We make use of freely available news summary data, transforming declarative sentences into appropriate questions using dependency parsing, named entity recognition and semantic role labeling.
The resulting questions are then combined with the original news articles to train an end-to-end neural QG model.
arXiv Detail & Related papers (2021-09-16T13:08:43Z) - NoiseQA: Challenge Set Evaluation for User-Centric Question Answering [68.67783808426292]
We show that components in the pipeline that precede an answering engine can introduce varied and considerable sources of error.
We conclude that there is substantial room for progress before QA systems can be effectively deployed.
arXiv Detail & Related papers (2021-02-16T18:35:29Z) - Knowledge-Routed Visual Question Reasoning: Challenges for Deep
Representation Embedding [140.5911760063681]
We propose a novel dataset named Knowledge-Routed Visual Question Reasoning for VQA model evaluation.
We generate the question-answer pair based on both the Visual Genome scene graph and an external knowledge base with controlled programs.
arXiv Detail & Related papers (2020-12-14T00:33:44Z) - The Impact of Explanations on AI Competency Prediction in VQA [3.149760860038061]
We evaluate the impact of explanations on the user's mental model of AI agent competency within the task of visual question answering (VQA)
We introduce an explainable VQA system that uses spatial and object features and is powered by the BERT language model.
arXiv Detail & Related papers (2020-07-02T06:11:28Z) - Review-guided Helpful Answer Identification in E-commerce [38.276241153439955]
Product-specific community question answering platforms can greatly help address the concerns of potential customers.
The user-provided answers on such platforms often vary a lot in their qualities.
Helpfulness votes from the community can indicate the overall quality of the answer, but they are often missing.
arXiv Detail & Related papers (2020-03-13T11:34:29Z) - SQuINTing at VQA Models: Introspecting VQA Models with Sub-Questions [66.86887670416193]
We show that state-of-the-art VQA models have comparable performance in answering perception and reasoning questions, but suffer from consistency problems.
To address this shortcoming, we propose an approach called Sub-Question-aware Network Tuning (SQuINT)
We show that SQuINT improves model consistency by 5%, also marginally improving performance on the Reasoning questions in VQA, while also displaying better attention maps.
arXiv Detail & Related papers (2020-01-20T01:02:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.