Event Knowledge Incorporation with Posterior Regularization for
Event-Centric Question Answering
- URL: http://arxiv.org/abs/2305.04522v1
- Date: Mon, 8 May 2023 07:45:12 GMT
- Title: Event Knowledge Incorporation with Posterior Regularization for
Event-Centric Question Answering
- Authors: Junru Lu, Gabriele Pergola, Lin Gui, Yulan He
- Abstract summary: We propose a strategy to incorporate event knowledge extracted from event trigger annotations via posterior regularization.
In particular, we define event-related knowledge constraints based on the event trigger annotations in the QA datasets.
We conduct experiments on two event-centric QA datasets, TORQUE and ESTER.
- Score: 32.03893317439898
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a simple yet effective strategy to incorporate event knowledge
extracted from event trigger annotations via posterior regularization to
improve the event reasoning capability of mainstream question-answering (QA)
models for event-centric QA. In particular, we define event-related knowledge
constraints based on the event trigger annotations in the QA datasets, and
subsequently use them to regularize the posterior answer output probabilities
from the backbone pre-trained language models used in the QA setting. We
explore two different posterior regularization strategies for extractive and
generative QA separately. For extractive QA, the sentence-level event knowledge
constraint is defined by assessing if a sentence contains an answer event or
not, which is later used to modify the answer span extraction probability. For
generative QA, the token-level event knowledge constraint is defined by
comparing the generated token from the backbone language model with the answer
event in order to introduce a reward or penalty term, which essentially adjusts
the answer generative probability indirectly. We conduct experiments on two
event-centric QA datasets, TORQUE and ESTER. The results show that our proposed
approach can effectively inject event knowledge into existing pre-trained
language models and achieves strong performance compared to existing QA models
in answer evaluation. Code and models can be found:
https://github.com/LuJunru/EventQAviaPR.
Related papers
- QASnowball: An Iterative Bootstrapping Framework for High-Quality
Question-Answering Data Generation [67.27999343730224]
We introduce an iterative bootstrapping framework for QA data augmentation (named QASnowball)
QASnowball can iteratively generate large-scale high-quality QA data based on a seed set of supervised examples.
We conduct experiments in the high-resource English scenario and the medium-resource Chinese scenario, and the experimental results show that the data generated by QASnowball can facilitate QA models.
arXiv Detail & Related papers (2023-09-19T05:20:36Z) - DAPrompt: Deterministic Assumption Prompt Learning for Event Causality
Identification [8.102227953905206]
Event Causality Identification (ECI) aims at determining whether there is a causal relation between two event mentions.
We propose a deterministic assumption prompt learning model, called DAPrompt, for the ECI task.
We use the probabilities of predicted events to evaluate the assumption rationality for the final event causality decision.
arXiv Detail & Related papers (2023-07-19T08:02:20Z) - Event Extraction as Question Generation and Answering [72.04433206754489]
Recent work on Event Extraction has reframed the task as Question Answering (QA)
We propose QGA-EE, which enables a Question Generation (QG) model to generate questions that incorporate rich contextual information instead of using fixed templates.
Experiments show that QGA-EE outperforms all prior single-task-based models on the ACE05 English dataset.
arXiv Detail & Related papers (2023-07-10T01:46:15Z) - Event-Centric Question Answering via Contrastive Learning and Invertible
Event Transformation [29.60817278635999]
We propose a novel QA model with contrastive learning and invertible event transformation, called TranCLR.
Our proposed model utilizes an invertible transformation matrix to project semantic vectors of events into a common event embedding space, trained with contrastive learning, and thus naturally inject event semantic knowledge into mainstream QA pipelines.
arXiv Detail & Related papers (2022-10-24T01:15:06Z) - Learning Constraints and Descriptive Segmentation for Subevent Detection [74.48201657623218]
We propose an approach to learning and enforcing constraints that capture dependencies between subevent detection and EventSeg prediction.
We adopt Rectifier Networks for constraint learning and then convert the learned constraints to a regularization term in the loss function of the neural model.
arXiv Detail & Related papers (2021-09-13T20:50:37Z) - Event Detection as Question Answering with Entity Information [5.761450181435801]
We propose a paradigm for the task of event detection (ED) by casting it as a question-answering (QA) problem with the possibility of multiple answers and the support of entities.
The extraction of event triggers is, thus, transformed into the task of identifying answer spans from a context, while also focusing on the surrounding entities.
Experiments on the ACE2005 corpus demonstrate that the proposed paradigm is a viable solution for the ED task and it significantly outperforms the state-of-the-art models.
arXiv Detail & Related papers (2021-04-14T16:53:11Z) - Detecting Ongoing Events Using Contextual Word and Sentence Embeddings [110.83289076967895]
This paper introduces the Ongoing Event Detection (OED) task.
The goal is to detect ongoing event mentions only, as opposed to historical, future, hypothetical, or other forms or events that are neither fresh nor current.
Any application that needs to extract structured information about ongoing events from unstructured texts can take advantage of an OED system.
arXiv Detail & Related papers (2020-07-02T20:44:05Z) - Harvesting and Refining Question-Answer Pairs for Unsupervised QA [95.9105154311491]
We introduce two approaches to improve unsupervised Question Answering (QA)
First, we harvest lexically and syntactically divergent questions from Wikipedia to automatically construct a corpus of question-answer pairs (named as RefQA)
Second, we take advantage of the QA model to extract more appropriate answers, which iteratively refines data over RefQA.
arXiv Detail & Related papers (2020-05-06T15:56:06Z) - Template-Based Question Generation from Retrieved Sentences for Improved
Unsupervised Question Answering [98.48363619128108]
We propose an unsupervised approach to training QA models with generated pseudo-training data.
We show that generating questions for QA training by applying a simple template on a related, retrieved sentence rather than the original context sentence improves downstream QA performance.
arXiv Detail & Related papers (2020-04-24T17:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.