An Ensemble Approach to Acronym Extraction using Transformers
- URL: http://arxiv.org/abs/2201.03026v1
- Date: Sun, 9 Jan 2022 14:49:46 GMT
- Title: An Ensemble Approach to Acronym Extraction using Transformers
- Authors: Prashant Sharma, Hadeel Saadany, Leonardo Zilio, Diptesh Kanojia,
Constantin Or\u{a}san
- Abstract summary: Acronyms are abbreviated units of a phrase constructed by using initial components of the phrase in a text.
This paper discusses an ensemble approach for the task of Acronym Extraction.
- Score: 7.88595796865485
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Acronyms are abbreviated units of a phrase constructed by using initial
components of the phrase in a text. Automatic extraction of acronyms from a
text can help various Natural Language Processing tasks like machine
translation, information retrieval, and text summarisation. This paper
discusses an ensemble approach for the task of Acronym Extraction, which
utilises two different methods to extract acronyms and their corresponding long
forms. The first method utilises a multilingual contextual language model and
fine-tunes the model to perform the task. The second method relies on a
convolutional neural network architecture to extract acronyms and append them
to the output of the previous method. We also augment the official training
dataset with additional training samples extracted from several open-access
journals to help improve the task performance. Our dataset analysis also
highlights the noise within the current task dataset. Our approach achieves the
following macro-F1 scores on test data released with the task: Danish (0.74),
English-Legal (0.72), English-Scientific (0.73), French (0.63), Persian (0.57),
Spanish (0.65), Vietnamese (0.65). We release our code and models publicly.
Related papers
- A Novel Cartography-Based Curriculum Learning Method Applied on RoNLI: The First Romanian Natural Language Inference Corpus [71.77214818319054]
Natural language inference is a proxy for natural language understanding.
There is no publicly available NLI corpus for the Romanian language.
We introduce the first Romanian NLI corpus (RoNLI) comprising 58K training sentence pairs.
arXiv Detail & Related papers (2024-05-20T08:41:15Z) - Cross-lingual Contextualized Phrase Retrieval [63.80154430930898]
We propose a new task formulation of dense retrieval, cross-lingual contextualized phrase retrieval.
We train our Cross-lingual Contextualized Phrase Retriever (CCPR) using contrastive learning.
On the phrase retrieval task, CCPR surpasses baselines by a significant margin, achieving a top-1 accuracy that is at least 13 points higher.
arXiv Detail & Related papers (2024-03-25T14:46:51Z) - Language and Task Arithmetic with Parameter-Efficient Layers for Zero-Shot Summarization [126.96113831681338]
In this paper, we propose to improve zero-shot cross-lingual transfer by composing language or task specialized parameters.
Our method composes language and task PEFT modules via element-wise arithmetic operations to leverage unlabeled data and English labeled data.
arXiv Detail & Related papers (2023-11-15T20:04:58Z) - Ensemble Transfer Learning for Multilingual Coreference Resolution [60.409789753164944]
A problem that frequently occurs when working with a non-English language is the scarcity of annotated training data.
We design a simple but effective ensemble-based framework that combines various transfer learning techniques.
We also propose a low-cost TL method that bootstraps coreference resolution models by utilizing Wikipedia anchor texts.
arXiv Detail & Related papers (2023-01-22T18:22:55Z) - Training Naturalized Semantic Parsers with Very Little Data [10.709587018625275]
State-of-the-art (SOTA) semantics are seq2seq architectures based on large language models that have been pretrained on vast amounts of text.
Recent work has explored a reformulation of semantic parsing whereby the output sequences are themselves natural language sentences.
We show that this method delivers new SOTA few-shot performance on the Overnight dataset.
arXiv Detail & Related papers (2022-04-29T17:14:54Z) - PLOD: An Abbreviation Detection Dataset for Scientific Documents [8.085950562565893]
PLOD is a large-scale dataset for abbreviation detection and extraction.
It contains 160k+ segments automatically annotated with abbreviations and their long forms.
We generate several baseline models for detecting abbreviations and long forms.
arXiv Detail & Related papers (2022-04-26T03:52:21Z) - PSG: Prompt-based Sequence Generation for Acronym Extraction [26.896811663334162]
We propose a Prompt-based Sequence Generation (PSG) method for the acronym extraction task.
Specifically, we design a template for prompting the extracted acronym texts with auto-regression.
A position extraction algorithm is designed for extracting the position of the generated answers.
arXiv Detail & Related papers (2021-11-29T02:14:38Z) - Zero-Shot Information Extraction as a Unified Text-to-Triple Translation [56.01830747416606]
We cast a suite of information extraction tasks into a text-to-triple translation framework.
We formalize the task as a translation between task-specific input text and output triples.
We study the zero-shot performance of this framework on open information extraction.
arXiv Detail & Related papers (2021-09-23T06:54:19Z) - MCL@IITK at SemEval-2021 Task 2: Multilingual and Cross-lingual
Word-in-Context Disambiguation using Augmented Data, Signals, and
Transformers [1.869621561196521]
We present our approach for solving the SemEval 2021 Task 2: Multilingual and Cross-lingual Word-in-Context Disambiguation (MCL-WiC)
The goal is to detect whether a given word common to both the sentences evokes the same meaning.
We submit systems for both the settings - Multilingual and Cross-Lingual.
arXiv Detail & Related papers (2021-04-04T08:49:28Z) - Pre-training via Paraphrasing [96.79972492585112]
We introduce MARGE, a pre-trained sequence-to-sequence model learned with an unsupervised multi-lingual paraphrasing objective.
We show it is possible to jointly learn to do retrieval and reconstruction, given only a random initialization.
For example, with no additional task-specific training we achieve BLEU scores of up to 35.8 for document translation.
arXiv Detail & Related papers (2020-06-26T14:43:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.