Consecutive Question Generation via Dynamic Multitask Learning
- URL: http://arxiv.org/abs/2211.08850v1
- Date: Wed, 16 Nov 2022 11:50:36 GMT
- Title: Consecutive Question Generation via Dynamic Multitask Learning
- Authors: Yunji Li, Sujian Li, Xing Shi
- Abstract summary: We propose the task of consecutive question generation (CQG), which generates a set of logically related question-answer pairs to understand a whole passage.
We first examine the four key elements of CQG, and propose a novel dynamic multitask framework with one main task generating a question-answer pair, and four auxiliary tasks generating other elements.
We prove that our strategy can improve question generation significantly and benefit multiple related NLP tasks.
- Score: 17.264399861776187
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we propose the task of consecutive question generation (CQG),
which generates a set of logically related question-answer pairs to understand
a whole passage, with a comprehensive consideration of the aspects including
accuracy, coverage, and informativeness. To achieve this, we first examine the
four key elements of CQG, i.e., question, answer, rationale, and context
history, and propose a novel dynamic multitask framework with one main task
generating a question-answer pair, and four auxiliary tasks generating other
elements. It directly helps the model generate good questions through both
joint training and self-reranking. At the same time, to fully explore the
worth-asking information in a given passage, we make use of the reranking
losses to sample the rationales and search for the best question series
globally. Finally, we measure our strategy by QA data augmentation and manual
evaluation, as well as a novel application of generated question-answer pairs
on DocNLI. We prove that our strategy can improve question generation
significantly and benefit multiple related NLP tasks.
Related papers
- An Empirical Comparison of LM-based Question and Answer Generation
Methods [79.31199020420827]
Question and answer generation (QAG) consists of generating a set of question-answer pairs given a context.
In this paper, we establish baselines with three different QAG methodologies that leverage sequence-to-sequence language model (LM) fine-tuning.
Experiments show that an end-to-end QAG model, which is computationally light at both training and inference times, is generally robust and outperforms other more convoluted approaches.
arXiv Detail & Related papers (2023-05-26T14:59:53Z) - Rethinking Label Smoothing on Multi-hop Question Answering [87.68071401870283]
Multi-Hop Question Answering (MHQA) is a significant area in question answering.
In this work, we analyze the primary factors limiting the performance of multi-hop reasoning.
We propose a novel label smoothing technique, F1 Smoothing, which incorporates uncertainty into the learning process.
arXiv Detail & Related papers (2022-12-19T14:48:08Z) - Modern Question Answering Datasets and Benchmarks: A Survey [5.026863544662493]
Question Answering (QA) is one of the most important natural language processing (NLP) tasks.
It aims using NLP technologies to generate a corresponding answer to a given question based on the massive unstructured corpus.
In this paper, we investigate influential QA datasets that have been released in the era of deep learning.
arXiv Detail & Related papers (2022-06-30T05:53:56Z) - What should I Ask: A Knowledge-driven Approach for Follow-up Questions
Generation in Conversational Surveys [63.51903260461746]
We propose a novel task for knowledge-driven follow-up question generation in conversational surveys.
We constructed a new human-annotated dataset of human-written follow-up questions with dialogue history and labeled knowledge.
We then propose a two-staged knowledge-driven model for the task, which generates informative and coherent follow-up questions.
arXiv Detail & Related papers (2022-05-23T00:57:33Z) - Asking Questions Like Educational Experts: Automatically Generating
Question-Answer Pairs on Real-World Examination Data [10.353009081072992]
This paper addresses the question-answer pair generation task on the real-world examination data, and proposes a new unified framework on RACE.
We propose a multi-agent communication model to generate and optimize the question and keyphrases iteratively, and then apply the generated question and keyphrases to guide the generation of answers.
Experimental results show that our model makes great breakthroughs in the question-answer pair generation task.
arXiv Detail & Related papers (2021-09-11T04:10:57Z) - Generating Self-Contained and Summary-Centric Question Answer Pairs via
Differentiable Reward Imitation Learning [7.2745835227138045]
We propose a model for generating question-answer pairs (QA pairs) with self-contained, summary-centric questions and length-constrained, article-summarizing answers.
This dataset is used to learn a QA pair generation model producing summaries as answers that balance brevity with sufficiency jointly with their corresponding questions.
arXiv Detail & Related papers (2021-09-10T06:34:55Z) - Enhancing Question Generation with Commonsense Knowledge [33.289599417096206]
We propose a multi-task learning framework to introduce commonsense knowledge into question generation process.
Experimental results on SQuAD show that our proposed methods are able to noticeably improve the QG performance on both automatic and human evaluation metrics.
arXiv Detail & Related papers (2021-06-19T08:58:13Z) - A Dataset of Information-Seeking Questions and Answers Anchored in
Research Papers [66.11048565324468]
We present a dataset of 5,049 questions over 1,585 Natural Language Processing papers.
Each question is written by an NLP practitioner who read only the title and abstract of the corresponding paper, and the question seeks information present in the full text.
We find that existing models that do well on other QA tasks do not perform well on answering these questions, underperforming humans by at least 27 F1 points when answering them from entire papers.
arXiv Detail & Related papers (2021-05-07T00:12:34Z) - Inquisitive Question Generation for High Level Text Comprehension [60.21497846332531]
We introduce INQUISITIVE, a dataset of 19K questions that are elicited while a person is reading through a document.
We show that readers engage in a series of pragmatic strategies to seek information.
We evaluate question generation models based on GPT-2 and show that our model is able to generate reasonable questions.
arXiv Detail & Related papers (2020-10-04T19:03:39Z) - Reinforced Multi-task Approach for Multi-hop Question Generation [47.15108724294234]
We take up Multi-hop question generation, which aims at generating relevant questions based on supporting facts in the context.
We employ multitask learning with the auxiliary task of answer-aware supporting fact prediction to guide the question generator.
We demonstrate the effectiveness of our approach through experiments on the multi-hop question answering dataset, HotPotQA.
arXiv Detail & Related papers (2020-04-05T10:16:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.