Ask Question First for Enhancing Lifelong Language Learning
- URL: http://arxiv.org/abs/2208.08367v1
- Date: Wed, 17 Aug 2022 15:58:33 GMT
- Title: Ask Question First for Enhancing Lifelong Language Learning
- Authors: Han Wang, Ruiliu Fu, Xuejun Zhang, Jun Zhou, Qingwei Zhao
- Abstract summary: Lifelong language learning aims to stream learning NLP tasks while retaining knowledge of previous tasks.
Previous works have explored formatting all data as "begin token (textitB) + context (textitC) + question (textitQ + answer (textitA)" for different tasks.
We propose the Ask Question First and Replay Question (AQF-RQ), including a novel data format "textitBQCA" and a new training task to train pseudo questions of previous tasks.
- Score: 12.766084031891815
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lifelong language learning aims to stream learning NLP tasks while retaining
knowledge of previous tasks. Previous works based on the language model and
following data-free constraint approaches have explored formatting all data as
"begin token (\textit{B}) + context (\textit{C}) + question (\textit{Q}) +
answer (\textit{A})" for different tasks. However, they still suffer from
catastrophic forgetting and are exacerbated when the previous task's pseudo
data is insufficient for the following reasons: (1) The model has difficulty
generating task-corresponding pseudo data, and (2) \textit{A} is prone to error
when \textit{A} and \textit{C} are separated by \textit{Q} because the
information of the \textit{C} is diminished before generating \textit{A}.
Therefore, we propose the Ask Question First and Replay Question (AQF-RQ),
including a novel data format "\textit{BQCA}" and a new training task to train
pseudo questions of previous tasks. Experimental results demonstrate that
AQF-RQ makes it easier for the model to generate more pseudo data that match
corresponding tasks, and is more robust to both sufficient and insufficient
pseudo-data when the task boundary is both clear and unclear. AQF-RQ can
achieve only 0.36\% lower performance than multi-task learning.
Related papers
- Event Extraction as Question Generation and Answering [72.04433206754489]
Recent work on Event Extraction has reframed the task as Question Answering (QA)
We propose QGA-EE, which enables a Question Generation (QG) model to generate questions that incorporate rich contextual information instead of using fixed templates.
Experiments show that QGA-EE outperforms all prior single-task-based models on the ACE05 English dataset.
arXiv Detail & Related papers (2023-07-10T01:46:15Z) - PAXQA: Generating Cross-lingual Question Answering Examples at Training
Scale [53.92008514395125]
PAXQA (Projecting annotations for cross-lingual (x) QA) decomposes cross-lingual QA into two stages.
We propose a novel use of lexically-constrained machine translation, in which constrained entities are extracted from the parallel bitexts.
We show that models fine-tuned on these datasets outperform prior synthetic data generation models over several extractive QA datasets.
arXiv Detail & Related papers (2023-04-24T15:46:26Z) - Toward Unsupervised Realistic Visual Question Answering [70.67698100148414]
We study the problem of realistic VQA (RVQA), where a model has to reject unanswerable questions (UQs) and answer answerable ones (AQs)
We first point out 2 drawbacks in current RVQA research, where (1) datasets contain too many unchallenging UQs and (2) a large number of annotated UQs are required for training.
We propose a new testing dataset, RGQA, which combines AQs from an existing VQA dataset with around 29K human-annotated UQs.
This combines pseudo UQs obtained by randomly pairing images and questions, with an
arXiv Detail & Related papers (2023-03-09T06:58:29Z) - Interleaving Retrieval with Chain-of-Thought Reasoning for
Knowledge-Intensive Multi-Step Questions [50.114651561111245]
We propose IRCoT, a new approach for multi-step question answering.
It interleaves retrieval with steps in a CoT, guiding the retrieval with CoT and in turn using retrieved results to improve CoT.
arXiv Detail & Related papers (2022-12-20T18:26:34Z) - Modern Question Answering Datasets and Benchmarks: A Survey [5.026863544662493]
Question Answering (QA) is one of the most important natural language processing (NLP) tasks.
It aims using NLP technologies to generate a corresponding answer to a given question based on the massive unstructured corpus.
In this paper, we investigate influential QA datasets that have been released in the era of deep learning.
arXiv Detail & Related papers (2022-06-30T05:53:56Z) - New Methods & Metrics for LFQA tasks [0.0]
Long-form question answering tasks require retrieving the documents pertinent to a query, using them to form a paragraph-length answer.
This work addresses the train/validation/test dataset overlap, absence of automatic metrics and generated answers not being "grounded" in retrieved documents.
arXiv Detail & Related papers (2021-12-26T18:38:05Z) - Understanding Unnatural Questions Improves Reasoning over Text [54.235828149899625]
Complex question answering (CQA) over raw text is a challenging task.
Learning an effective CQA model requires large amounts of human-annotated data.
We address the challenge of learning a high-quality programmer (parser) by projecting natural human-generated questions into unnatural machine-generated questions.
arXiv Detail & Related papers (2020-10-19T10:22:16Z) - Learning to Explain: Datasets and Models for Identifying Valid Reasoning
Chains in Multihop Question-Answering [28.67167530758428]
We introduce three datasets in which explanations formed from corpus facts are annotated.
eQASC contains over 98K explanation annotations for the multihop question answering dataset QASC.
eQASC-perturbed is constructed by crowd-sourcing perturbations to test consistency and generalization of explanation prediction models.
eOBQA is constructed by adding explanation annotations to the OBQA dataset to test generalization of models trained on eQASC.
arXiv Detail & Related papers (2020-10-07T08:46:02Z) - Tell Me How to Ask Again: Question Data Augmentation with Controllable
Rewriting in Continuous Space [94.8320535537798]
Controllable Rewriting based Question Data Augmentation (CRQDA) for machine reading comprehension (MRC), question generation, and question-answering natural language inference tasks.
We treat the question data augmentation task as a constrained question rewriting problem to generate context-relevant, high-quality, and diverse question data samples.
arXiv Detail & Related papers (2020-10-04T03:13:46Z) - Harvesting and Refining Question-Answer Pairs for Unsupervised QA [95.9105154311491]
We introduce two approaches to improve unsupervised Question Answering (QA)
First, we harvest lexically and syntactically divergent questions from Wikipedia to automatically construct a corpus of question-answer pairs (named as RefQA)
Second, we take advantage of the QA model to extract more appropriate answers, which iteratively refines data over RefQA.
arXiv Detail & Related papers (2020-05-06T15:56:06Z) - Template-Based Question Generation from Retrieved Sentences for Improved
Unsupervised Question Answering [98.48363619128108]
We propose an unsupervised approach to training QA models with generated pseudo-training data.
We show that generating questions for QA training by applying a simple template on a related, retrieved sentence rather than the original context sentence improves downstream QA performance.
arXiv Detail & Related papers (2020-04-24T17:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.