Diversifying Question Generation over Knowledge Base via External
Natural Questions
- URL: http://arxiv.org/abs/2309.14362v1
- Date: Sat, 23 Sep 2023 10:37:57 GMT
- Title: Diversifying Question Generation over Knowledge Base via External
Natural Questions
- Authors: Shasha Guo, Jing Zhang, Xirui Ke, Cuiping Li, Hong Chen
- Abstract summary: We argue that diverse texts should convey the same semantics through varied expressions.
Current metrics inadequately assess the above diversity since they calculate the ratio of unique n-grams in the generated question itself.
We devise a new diversity evaluation metric, which measures the diversity among top-k generated questions for each instance.
- Score: 18.382095354733842
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Previous methods on knowledge base question generation (KBQG) primarily focus
on enhancing the quality of a single generated question. Recognizing the
remarkable paraphrasing ability of humans, we contend that diverse texts should
convey the same semantics through varied expressions. The above insights make
diversifying question generation an intriguing task, where the first challenge
is evaluation metrics for diversity. Current metrics inadequately assess the
above diversity since they calculate the ratio of unique n-grams in the
generated question itself, which leans more towards measuring duplication
rather than true diversity. Accordingly, we devise a new diversity evaluation
metric, which measures the diversity among top-k generated questions for each
instance while ensuring their relevance to the ground truth. Clearly, the
second challenge is how to enhance diversifying question generation. To address
this challenge, we introduce a dual model framework interwoven by two selection
strategies to generate diverse questions leveraging external natural questions.
The main idea of our dual framework is to extract more diverse expressions and
integrate them into the generation model to enhance diversifying question
generation. Extensive experiments on widely used benchmarks for KBQG
demonstrate that our proposed approach generates highly diverse questions and
improves the performance of question answering tasks.
Related papers
- Qsnail: A Questionnaire Dataset for Sequential Question Generation [76.616068047362]
We present the first dataset specifically constructed for the questionnaire generation task, which comprises 13,168 human-written questionnaires.
We conduct experiments on Qsnail, and the results reveal that retrieval models and traditional generative models do not fully align with the given research topic and intents.
Despite enhancements through the chain-of-thought prompt and finetuning, questionnaires generated by language models still fall short of human-written questionnaires.
arXiv Detail & Related papers (2024-02-22T04:14:10Z) - Diversity Enhanced Narrative Question Generation for Storybooks [4.043005183192124]
We introduce a multi-question generation model (mQG) capable of generating multiple, diverse, and answerable questions.
To validate the answerability of the generated questions, we employ a SQuAD2.0 fine-tuned question answering model.
mQG shows promising results across various evaluation metrics, among strong baselines.
arXiv Detail & Related papers (2023-10-25T08:10:04Z) - Diversify Question Generation with Retrieval-Augmented Style Transfer [68.00794669873196]
We propose RAST, a framework for Retrieval-Augmented Style Transfer.
The objective is to utilize the style of diverse templates for question generation.
We develop a novel Reinforcement Learning (RL) based approach that maximizes a weighted combination of diversity reward and consistency reward.
arXiv Detail & Related papers (2023-10-23T02:27:31Z) - Improving Question Generation with Multi-level Content Planning [70.37285816596527]
This paper addresses the problem of generating questions from a given context and an answer, specifically focusing on questions that require multi-hop reasoning across an extended context.
We propose MultiFactor, a novel QG framework based on multi-level content planning. Specifically, MultiFactor includes two components: FA-model, which simultaneously selects key phrases and generates full answers, and Q-model which takes the generated full answer as an additional input to generate questions.
arXiv Detail & Related papers (2023-10-20T13:57:01Z) - Towards Diverse and Effective Question-Answer Pair Generation from
Children Storybooks [3.850557558248366]
We propose a framework that enhances QA type diversity by producing different interrogative sentences and implicit/explicit answers.
Our framework comprises a QFS-based answer generator, an iterative QA generator, and a relevancy-aware ranker.
arXiv Detail & Related papers (2023-06-11T06:55:59Z) - Enhancing Question Generation with Commonsense Knowledge [33.289599417096206]
We propose a multi-task learning framework to introduce commonsense knowledge into question generation process.
Experimental results on SQuAD show that our proposed methods are able to noticeably improve the QG performance on both automatic and human evaluation metrics.
arXiv Detail & Related papers (2021-06-19T08:58:13Z) - Inquisitive Question Generation for High Level Text Comprehension [60.21497846332531]
We introduce INQUISITIVE, a dataset of 19K questions that are elicited while a person is reading through a document.
We show that readers engage in a series of pragmatic strategies to seek information.
We evaluate question generation models based on GPT-2 and show that our model is able to generate reasonable questions.
arXiv Detail & Related papers (2020-10-04T19:03:39Z) - Evaluating for Diversity in Question Generation over Text [5.369031521471668]
We argue that commonly-used evaluation metrics such as BLEU and METEOR are not suitable for this task due to the inherent diversity of reference questions.
We propose a variational encoder-decoder model for this task.
arXiv Detail & Related papers (2020-08-17T13:16:12Z) - Match$^2$: A Matching over Matching Model for Similar Question
Identification [74.7142127303489]
Community Question Answering (CQA) has become a primary means for people to acquire knowledge, where people are free to ask questions or submit answers.
Similar question identification becomes a core task in CQA which aims to find a similar question from the archived repository whenever a new question is asked.
It has long been a challenge to properly measure the similarity between two questions due to the inherent variation of natural language, i.e., there could be different ways to ask a same question or different questions sharing similar expressions.
Traditional methods typically take a one-side usage, which leverages the answer as some expanded representation of the
arXiv Detail & Related papers (2020-06-21T05:59:34Z) - Reinforced Multi-task Approach for Multi-hop Question Generation [47.15108724294234]
We take up Multi-hop question generation, which aims at generating relevant questions based on supporting facts in the context.
We employ multitask learning with the auxiliary task of answer-aware supporting fact prediction to guide the question generator.
We demonstrate the effectiveness of our approach through experiments on the multi-hop question answering dataset, HotPotQA.
arXiv Detail & Related papers (2020-04-05T10:16:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.