Textual Entailment for Event Argument Extraction: Zero- and Few-Shot
with Multi-Source Learning
- URL: http://arxiv.org/abs/2205.01376v1
- Date: Tue, 3 May 2022 08:53:55 GMT
- Title: Textual Entailment for Event Argument Extraction: Zero- and Few-Shot
with Multi-Source Learning
- Authors: Oscar Sainz, Itziar Gonzalez-Dios, Oier Lopez de Lacalle, Bonan Min
and Eneko Agirre
- Abstract summary: Recent work has shown that NLP tasks can be recasted as Textual Entailment tasks using verbalizations.
We show that entailment is also effective in Event Argument Extraction (EAE), reducing the need of manual annotation to 50% and 20%.
- Score: 22.531385318852426
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Recent work has shown that NLP tasks such as Relation Extraction (RE) can be
recasted as Textual Entailment tasks using verbalizations, with strong
performance in zero-shot and few-shot settings thanks to pre-trained entailment
models. The fact that relations in current RE datasets are easily verbalized
casts doubts on whether entailment would be effective in more complex tasks. In
this work we show that entailment is also effective in Event Argument
Extraction (EAE), reducing the need of manual annotation to 50% and 20% in ACE
and WikiEvents respectively, while achieving the same performance as with full
training. More importantly, we show that recasting EAE as entailment alleviates
the dependency on schemas, which has been a road-block for transferring
annotations between domains. Thanks to the entailment, the multi-source
transfer between ACE and WikiEvents further reduces annotation down to 10% and
5% (respectively) of the full training without transfer. Our analysis shows
that the key to good results is the use of several entailment datasets to
pre-train the entailment model. Similar to previous approaches, our method
requires a small amount of effort for manual verbalization: only less than 15
minutes per event argument type is needed, and comparable results can be
achieved with users with different level of expertise.
Related papers
- Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - Divide, Conquer, and Combine: Mixture of Semantic-Independent Experts
for Zero-Shot Dialogue State Tracking [83.40120598637665]
Zero-shot transfer learning for Dialogue State Tracking (DST) helps to handle a variety of task-oriented dialogue domains without the cost of collecting in-domain data.
Existing works mainly study common data- or model-level augmentation methods to enhance the generalization.
We present a simple and effective "divide, conquer and combine" solution, which explicitly disentangles the semantics of seen data.
arXiv Detail & Related papers (2023-06-01T08:21:20Z) - Silver Syntax Pre-training for Cross-Domain Relation Extraction [20.603482820770356]
Relation Extraction (RE) remains a challenging task, especially when considering realistic out-of-domain evaluations.
obtaining high-quality (manually annotated) data is extremely expensive and cannot realistically be repeated for each new domain.
An intermediate training step on data from related tasks has shown to be beneficial across many NLP tasks.However, this setup still requires supplementary annotated data, which is often not available.
In this paper, we investigate intermediate pre-training specifically for RE. We exploit the affinity between syntactic structure and semantic RE, and identify the syntactic relations closely related to RE by being on the shortest dependency path between two entities
arXiv Detail & Related papers (2023-05-18T14:49:19Z) - Can You Label Less by Using Out-of-Domain Data? Active & Transfer
Learning with Few-shot Instructions [58.69255121795761]
We propose a novel Active Transfer Few-shot Instructions (ATF) approach which requires no fine-tuning.
ATF leverages the internal linguistic knowledge of pre-trained language models (PLMs) to facilitate the transfer of information.
We show that annotation of just a few target-domain samples via active learning can be beneficial for transfer, but the impact diminishes with more annotation effort.
arXiv Detail & Related papers (2022-11-21T19:03:31Z) - Beyond Transfer Learning: Co-finetuning for Action Localisation [64.07196901012153]
We propose co-finetuning -- simultaneously training a single model on multiple upstream'' and downstream'' tasks.
We demonstrate that co-finetuning outperforms traditional transfer learning when using the same total amount of data.
We also show how we can easily extend our approach to multiple upstream'' datasets to further improve performance.
arXiv Detail & Related papers (2022-07-08T10:25:47Z) - Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than
In-Context Learning [81.3514358542452]
Few-shot in-context learning (ICL) incurs substantial computational, memory, and storage costs because it involves processing all of the training examples every time a prediction is made.
parameter-efficient fine-tuning offers an alternative paradigm where a small set of parameters are trained to enable a model to perform the new task.
In this paper, we rigorously compare few-shot ICL and parameter-efficient fine-tuning and demonstrate that the latter offers better accuracy as well as dramatically lower computational costs.
arXiv Detail & Related papers (2022-05-11T17:10:41Z) - PERFECT: Prompt-free and Efficient Few-shot Learning with Language
Models [67.3725459417758]
PERFECT is a simple and efficient method for few-shot fine-tuning of PLMs without relying on any such handcrafting.
We show that manually engineered task prompts can be replaced with task-specific adapters that enable sample-efficient fine-tuning.
Experiments on a wide range of few-shot NLP tasks demonstrate that PERFECT, while being simple and efficient, also outperforms existing state-of-the-art few-shot learning methods.
arXiv Detail & Related papers (2022-04-03T22:31:25Z) - Label Verbalization and Entailment for Effective Zero- and Few-Shot
Relation Extraction [25.151235270747804]
In this work we reformulate relation extraction as an entailment task, with simple, hand-made, verbalizations of relations produced in less than 15 min per relation.
The system relies on a pretrained textual entailment engine which is run as-is (no training examples, zero-shot) or further fine-tuned on labeled examples (few-shot or fully trained)
In experiments on TACRED we attain 63% F1 zero-shot, 69% with 16 examples per relation (17% points better than the best supervised system on the same conditions), and only 4 points short to the state-of-the
arXiv Detail & Related papers (2021-09-08T14:04:50Z) - Zero-Resource Multi-Dialectal Arabic Natural Language Understanding [0.0]
We investigate the zero-shot performance on Dialectal Arabic (DA) when fine-tuning a pre-trained language model on modern standard Arabic (MSA) data only.
We propose self-training with unlabeled DA data and apply it in the context of named entity recognition (NER), part-of-speech (POS) tagging, and sarcasm detection (SRD)
Our results demonstrate the effectiveness of self-training with unlabeled DA data.
arXiv Detail & Related papers (2021-04-14T02:29:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.