KEPR: Knowledge Enhancement and Plausibility Ranking for Generative
Commonsense Question Answering
- URL: http://arxiv.org/abs/2305.08347v1
- Date: Mon, 15 May 2023 04:58:37 GMT
- Title: KEPR: Knowledge Enhancement and Plausibility Ranking for Generative
Commonsense Question Answering
- Authors: Zhifeng Li and Bowei Zou and Yifan Fan and Yu Hong
- Abstract summary: We propose a Knowledge Enhancement and Plausibility Ranking approach grounded on the Generate-Then-Rank pipeline architecture.
Specifically, we expand questions in terms of Wiktionary commonsense knowledge of keywords, and reformulate them with normalized patterns.
We develop an ELECTRA-based answer ranking model, where logistic regression is conducted during training, with the aim of approxing different levels of plausibility.
- Score: 11.537283115693432
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative commonsense question answering (GenCQA) is a task of automatically
generating a list of answers given a question. The answer list is required to
cover all reasonable answers. This presents the considerable challenges of
producing diverse answers and ranking them properly. Incorporating a variety of
closely-related background knowledge into the encoding of questions enables the
generation of different answers. Meanwhile, learning to distinguish positive
answers from negative ones potentially enhances the probabilistic estimation of
plausibility, and accordingly, the plausibility-based ranking. Therefore, we
propose a Knowledge Enhancement and Plausibility Ranking (KEPR) approach
grounded on the Generate-Then-Rank pipeline architecture. Specifically, we
expand questions in terms of Wiktionary commonsense knowledge of keywords, and
reformulate them with normalized patterns. Dense passage retrieval is utilized
for capturing relevant knowledge, and different PLM-based (BART, GPT2 and T5)
networks are used for generating answers. On the other hand, we develop an
ELECTRA-based answer ranking model, where logistic regression is conducted
during training, with the aim of approximating different levels of plausibility
in a polar classification scenario. Extensive experiments on the benchmark
ProtoQA show that KEPR obtains substantial improvements, compared to the strong
baselines. Within the experimental models, the T5-based GenCQA with KEPR
obtains the best performance, which is up to 60.91% at the primary canonical
metric Inc@3. It outperforms the existing GenCQA models on the current
leaderboard of ProtoQA.
Related papers
- Prompting Large Language Models with Chain-of-Thought for Few-Shot
Knowledge Base Question Generation [19.327008532572645]
Question Generation over Knowledge Bases (KBQG) aims to convert a logical form into a natural language question.
We propose Chain-of-Thought prompting, which is an in-context learning strategy for reasoning.
We conduct extensive experiments over three public KBQG datasets.
arXiv Detail & Related papers (2023-10-12T15:08:14Z) - Generate then Select: Open-ended Visual Question Answering Guided by
World Knowledge [155.81786738036578]
Open-ended Visual Question Answering (VQA) task requires AI models to jointly reason over visual and natural language inputs.
Pre-trained Language Models (PLM) such as GPT-3 have been applied to the task and shown to be powerful world knowledge sources.
We propose RASO: a new VQA pipeline that deploys a generate-then-select strategy guided by world knowledge.
arXiv Detail & Related papers (2023-05-30T08:34:13Z) - An Empirical Comparison of LM-based Question and Answer Generation
Methods [79.31199020420827]
Question and answer generation (QAG) consists of generating a set of question-answer pairs given a context.
In this paper, we establish baselines with three different QAG methodologies that leverage sequence-to-sequence language model (LM) fine-tuning.
Experiments show that an end-to-end QAG model, which is computationally light at both training and inference times, is generally robust and outperforms other more convoluted approaches.
arXiv Detail & Related papers (2023-05-26T14:59:53Z) - Knowledge Transfer from Answer Ranking to Answer Generation [97.38378660163414]
We propose to train a GenQA model by transferring knowledge from a trained AS2 model.
We also propose to use the AS2 model prediction scores for loss weighting and score-conditioned input/output shaping.
arXiv Detail & Related papers (2022-10-23T21:51:27Z) - Answer Generation for Questions With Multiple Information Sources in
E-Commerce [0.0]
We propose a novel pipeline (MSQAP) that utilizes the rich information present in the aforementioned sources by separately performing relevancy and ambiguity prediction.
This is the first work in the e-commerce domain that automatically generates natural language answers combining the information present in diverse sources such as specifications, similar questions, and reviews data.
arXiv Detail & Related papers (2021-11-27T23:19:49Z) - Calculating Question Similarity is Enough:A New Method for KBQA Tasks [8.056701645706404]
This paper proposes a Corpus Generation - Retrieve Method (CGRM) with Pre-training Language Model (PLM) and Knowledge Graph (KG)
Firstly, based on the mT5 model, we designed two new pre-training tasks: knowledge masked language modeling and question generation based on the paragraph.
Secondly, after preprocessing triples of knowledge graph with a series of rules, the kT5 model generates natural language QA pairs based on processed triples.
arXiv Detail & Related papers (2021-11-15T10:31:46Z) - RnG-KBQA: Generation Augmented Iterative Ranking for Knowledge Base
Question Answering [57.94658176442027]
We present RnG-KBQA, a Rank-and-Generate approach for KBQA.
We achieve new state-of-the-art results on GrailQA and WebQSP datasets.
arXiv Detail & Related papers (2021-09-17T17:58:28Z) - Exploring Question-Specific Rewards for Generating Deep Questions [42.243227323241584]
We design three different rewards that target to improve the fluency, relevance, and answerability of generated questions.
We find that optimizing question-specific rewards generally leads to better performance in automatic evaluation metrics.
arXiv Detail & Related papers (2020-11-02T16:37:30Z) - Few-Shot Complex Knowledge Base Question Answering via Meta
Reinforcement Learning [55.08037694027792]
Complex question-answering (CQA) involves answering complex natural-language questions on a knowledge base (KB)
The conventional neural program induction (NPI) approach exhibits uneven performance when the questions have different types.
This paper proposes a meta-reinforcement learning approach to program induction in CQA to tackle the potential distributional bias in questions.
arXiv Detail & Related papers (2020-10-29T18:34:55Z) - PRover: Proof Generation for Interpretable Reasoning over Rules [81.40404921232192]
We propose a transformer-based model that answers binary questions over rule-bases and generates the corresponding proofs.
Our model learns to predict nodes and edges corresponding to proof graphs in an efficient constrained training paradigm.
We conduct experiments on synthetic, hand-authored, and human-paraphrased rule-bases to show promising results for QA and proof generation.
arXiv Detail & Related papers (2020-10-06T15:47:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.