Harvesting and Refining Question-Answer Pairs for Unsupervised QA
- URL: http://arxiv.org/abs/2005.02925v1
- Date: Wed, 6 May 2020 15:56:06 GMT
- Title: Harvesting and Refining Question-Answer Pairs for Unsupervised QA
- Authors: Zhongli Li, Wenhui Wang, Li Dong, Furu Wei, Ke Xu
- Abstract summary: We introduce two approaches to improve unsupervised Question Answering (QA)
First, we harvest lexically and syntactically divergent questions from Wikipedia to automatically construct a corpus of question-answer pairs (named as RefQA)
Second, we take advantage of the QA model to extract more appropriate answers, which iteratively refines data over RefQA.
- Score: 95.9105154311491
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Question Answering (QA) has shown great success thanks to the availability of
large-scale datasets and the effectiveness of neural models. Recent research
works have attempted to extend these successes to the settings with few or no
labeled data available. In this work, we introduce two approaches to improve
unsupervised QA. First, we harvest lexically and syntactically divergent
questions from Wikipedia to automatically construct a corpus of question-answer
pairs (named as RefQA). Second, we take advantage of the QA model to extract
more appropriate answers, which iteratively refines data over RefQA. We conduct
experiments on SQuAD 1.1, and NewsQA by fine-tuning BERT without access to
manually annotated data. Our approach outperforms previous unsupervised
approaches by a large margin and is competitive with early supervised models.
We also show the effectiveness of our approach in the few-shot learning
setting.
Related papers
- A Lightweight Method to Generate Unanswerable Questions in English [18.323248259867356]
We examine a simpler data augmentation method for unanswerable question generation in English.
We perform antonym and entity swaps on answerable questions.
Compared to the prior state-of-the-art, data generated with our training-free and lightweight strategy results in better models.
arXiv Detail & Related papers (2023-10-30T10:14:52Z) - QASnowball: An Iterative Bootstrapping Framework for High-Quality
Question-Answering Data Generation [67.27999343730224]
We introduce an iterative bootstrapping framework for QA data augmentation (named QASnowball)
QASnowball can iteratively generate large-scale high-quality QA data based on a seed set of supervised examples.
We conduct experiments in the high-resource English scenario and the medium-resource Chinese scenario, and the experimental results show that the data generated by QASnowball can facilitate QA models.
arXiv Detail & Related papers (2023-09-19T05:20:36Z) - Toward Unsupervised Realistic Visual Question Answering [70.67698100148414]
We study the problem of realistic VQA (RVQA), where a model has to reject unanswerable questions (UQs) and answer answerable ones (AQs)
We first point out 2 drawbacks in current RVQA research, where (1) datasets contain too many unchallenging UQs and (2) a large number of annotated UQs are required for training.
We propose a new testing dataset, RGQA, which combines AQs from an existing VQA dataset with around 29K human-annotated UQs.
This combines pseudo UQs obtained by randomly pairing images and questions, with an
arXiv Detail & Related papers (2023-03-09T06:58:29Z) - Closed-book Question Generation via Contrastive Learning [20.644215991166895]
We propose a new QG model empowered by a contrastive learning module and an answer reconstruction module.
We show how to leverage the proposed model to improve existing closed-book QA systems.
arXiv Detail & Related papers (2022-10-13T06:45:46Z) - Improving Unsupervised Question Answering via Summarization-Informed
Question Generation [47.96911338198302]
Question Generation (QG) is the task of generating a plausible question for a passage, answer> pair.
We make use of freely available news summary data, transforming declarative sentences into appropriate questions using dependency parsing, named entity recognition and semantic role labeling.
The resulting questions are then combined with the original news articles to train an end-to-end neural QG model.
arXiv Detail & Related papers (2021-09-16T13:08:43Z) - Generating Diverse and Consistent QA pairs from Contexts with
Information-Maximizing Hierarchical Conditional VAEs [62.71505254770827]
We propose a conditional variational autoencoder (HCVAE) for generating QA pairs given unstructured texts as contexts.
Our model obtains impressive performance gains over all baselines on both tasks, using only a fraction of data for training.
arXiv Detail & Related papers (2020-05-28T08:26:06Z) - Template-Based Question Generation from Retrieved Sentences for Improved
Unsupervised Question Answering [98.48363619128108]
We propose an unsupervised approach to training QA models with generated pseudo-training data.
We show that generating questions for QA training by applying a simple template on a related, retrieved sentence rather than the original context sentence improves downstream QA performance.
arXiv Detail & Related papers (2020-04-24T17:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.