Ask To The Point: Open-Domain Entity-Centric Question Generation
- URL: http://arxiv.org/abs/2310.14126v1
- Date: Sat, 21 Oct 2023 22:19:19 GMT
- Title: Ask To The Point: Open-Domain Entity-Centric Question Generation
- Authors: Yuxiang Liu, Jie Huang, Kevin Chen-Chuan Chang
- Abstract summary: We introduce a new task called *entity-centric question generation* (ECQG)
The task aims to generate questions from an entity perspective.
To solve ECQG, we propose a coherent PLM-based framework GenCONE with two novel modules: content focusing and question verification.
- Score: 27.5948850672624
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce a new task called *entity-centric question generation* (ECQG),
motivated by real-world applications such as topic-specific learning, assisted
reading, and fact-checking. The task aims to generate questions from an entity
perspective. To solve ECQG, we propose a coherent PLM-based framework GenCONE
with two novel modules: content focusing and question verification. The content
focusing module first identifies a focus as "what to ask" to form draft
questions, and the question verification module refines the questions
afterwards by verifying the answerability. We also construct a large-scale
open-domain dataset from SQuAD to support this task. Our extensive experiments
demonstrate that GenCONE significantly and consistently outperforms various
baselines, and two modules are effective and complementary in generating
high-quality questions.
Related papers
- Benchmarking Multimodal Retrieval Augmented Generation with Dynamic VQA Dataset and Self-adaptive Planning Agent [102.31558123570437]
Multimodal Retrieval Augmented Generation (mRAG) plays an important role in mitigating the "hallucination" issue inherent in multimodal large language models (MLLMs)
We propose the first self-adaptive planning agent for multimodal retrieval, OmniSearch.
arXiv Detail & Related papers (2024-11-05T09:27:21Z) - Improving Question Generation with Multi-level Content Planning [70.37285816596527]
This paper addresses the problem of generating questions from a given context and an answer, specifically focusing on questions that require multi-hop reasoning across an extended context.
We propose MultiFactor, a novel QG framework based on multi-level content planning. Specifically, MultiFactor includes two components: FA-model, which simultaneously selects key phrases and generates full answers, and Q-model which takes the generated full answer as an additional input to generate questions.
arXiv Detail & Related papers (2023-10-20T13:57:01Z) - Consecutive Question Generation via Dynamic Multitask Learning [17.264399861776187]
We propose the task of consecutive question generation (CQG), which generates a set of logically related question-answer pairs to understand a whole passage.
We first examine the four key elements of CQG, and propose a novel dynamic multitask framework with one main task generating a question-answer pair, and four auxiliary tasks generating other elements.
We prove that our strategy can improve question generation significantly and benefit multiple related NLP tasks.
arXiv Detail & Related papers (2022-11-16T11:50:36Z) - Modern Question Answering Datasets and Benchmarks: A Survey [5.026863544662493]
Question Answering (QA) is one of the most important natural language processing (NLP) tasks.
It aims using NLP technologies to generate a corresponding answer to a given question based on the massive unstructured corpus.
In this paper, we investigate influential QA datasets that have been released in the era of deep learning.
arXiv Detail & Related papers (2022-06-30T05:53:56Z) - Multifaceted Improvements for Conversational Open-Domain Question
Answering [54.913313912927045]
We propose a framework with Multifaceted Improvements for Conversational open-domain Question Answering (MICQA)
Firstly, the proposed KL-divergence based regularization is able to lead to a better question understanding for retrieval and answer reading.
Second, the added post-ranker module can push more relevant passages to the top placements and be selected for reader with a two-aspect constrains.
Third, the well designed curriculum learning strategy effectively narrows the gap between the golden passage settings of training and inference, and encourages the reader to find true answer without the golden passage assistance.
arXiv Detail & Related papers (2022-04-01T07:54:27Z) - An Improved Attention for Visual Question Answering [46.89101543660587]
We consider the problem of Visual Question Answering (VQA)
Given an image and a free-form, open-ended, question, expressed in natural language, the goal of VQA system is to provide accurate answer to this question with respect to the image.
Attention, which captures intra- and inter-modal dependencies, has emerged as perhaps the most widely used mechanism for addressing these challenges.
arXiv Detail & Related papers (2020-11-04T07:34:54Z) - Inquisitive Question Generation for High Level Text Comprehension [60.21497846332531]
We introduce INQUISITIVE, a dataset of 19K questions that are elicited while a person is reading through a document.
We show that readers engage in a series of pragmatic strategies to seek information.
We evaluate question generation models based on GPT-2 and show that our model is able to generate reasonable questions.
arXiv Detail & Related papers (2020-10-04T19:03:39Z) - Tell Me How to Ask Again: Question Data Augmentation with Controllable
Rewriting in Continuous Space [94.8320535537798]
Controllable Rewriting based Question Data Augmentation (CRQDA) for machine reading comprehension (MRC), question generation, and question-answering natural language inference tasks.
We treat the question data augmentation task as a constrained question rewriting problem to generate context-relevant, high-quality, and diverse question data samples.
arXiv Detail & Related papers (2020-10-04T03:13:46Z) - ClarQ: A large-scale and diverse dataset for Clarification Question
Generation [67.1162903046619]
We devise a novel bootstrapping framework that assists in the creation of a diverse, large-scale dataset of clarification questions based on postcomments extracted from stackexchange.
We quantitatively demonstrate the utility of the newly created dataset by applying it to the downstream task of question-answering.
We release this dataset in order to foster research into the field of clarification question generation with the larger goal of enhancing dialog and question answering systems.
arXiv Detail & Related papers (2020-06-10T17:56:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.