Primer AI's Systems for Acronym Identification and Disambiguation
- URL: http://arxiv.org/abs/2012.08013v2
- Date: Wed, 6 Jan 2021 00:50:40 GMT
- Title: Primer AI's Systems for Acronym Identification and Disambiguation
- Authors: Nicholas Egan, John Bohannon
- Abstract summary: We introduce new methods for acronym identification and disambiguation.
Our systems achieve significant performance gains over previously suggested methods.
Both of our systems perform competitively on the SDU@AAAI-21 shared task leaderboard.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The prevalence of ambiguous acronyms make scientific documents harder to
understand for humans and machines alike, presenting a need for models that can
automatically identify acronyms in text and disambiguate their meaning. We
introduce new methods for acronym identification and disambiguation: our
acronym identification model projects learned token embeddings onto tag
predictions, and our acronym disambiguation model finds training examples with
similar sentence embeddings as test examples. Both of our systems achieve
significant performance gains over previously suggested methods, and perform
competitively on the SDU@AAAI-21 shared task leaderboard. Our models were
trained in part on new distantly-supervised datasets for these tasks which we
call AuxAI and AuxAD. We also identified a duplication conflict issue in the
SciAD dataset, and formed a deduplicated version of SciAD that we call
SciAD-dedupe. We publicly released all three of these datasets, and hope that
they help the community make further strides in scientific document
understanding.
Related papers
- Language Models As Semantic Indexers [78.83425357657026]
We introduce LMIndexer, a self-supervised framework to learn semantic IDs with a generative language model.
We show the high quality of the learned IDs and demonstrate their effectiveness on three tasks including recommendation, product search, and document retrieval.
arXiv Detail & Related papers (2023-10-11T18:56:15Z) - Unsupervised Sentiment Analysis of Plastic Surgery Social Media Posts [91.3755431537592]
The massive collection of user posts across social media platforms is primarily untapped for artificial intelligence (AI) use cases.
Natural language processing (NLP) is a subfield of AI that leverages bodies of documents, known as corpora, to train computers in human-like language understanding.
This study demonstrates that the applied results of unsupervised analysis allow a computer to predict either negative, positive, or neutral user sentiment towards plastic surgery.
arXiv Detail & Related papers (2023-07-05T20:16:20Z) - Disambiguation of Company names via Deep Recurrent Networks [101.90357454833845]
We propose a Siamese LSTM Network approach to extract -- via supervised learning -- an embedding of company name strings.
We analyse how an Active Learning approach to prioritise the samples to be labelled leads to a more efficient overall learning pipeline.
arXiv Detail & Related papers (2023-03-07T15:07:57Z) - DetCLIP: Dictionary-Enriched Visual-Concept Paralleled Pre-training for
Open-world Detection [118.36746273425354]
This paper presents a paralleled visual-concept pre-training method for open-world detection by resorting to knowledge enrichment from a designed concept dictionary.
By enriching the concepts with their descriptions, we explicitly build the relationships among various concepts to facilitate the open-domain learning.
The proposed framework demonstrates strong zero-shot detection performances, e.g., on the LVIS dataset, our DetCLIP-T outperforms GLIP-T by 9.9% mAP and obtains a 13.5% improvement on rare categories.
arXiv Detail & Related papers (2022-09-20T02:01:01Z) - BERT-based Acronym Disambiguation with Multiple Training Strategies [8.82012912690778]
Acronym disambiguation (AD) task aims to find the correct expansions of an ambiguous ancronym in a given sentence.
We propose a binary classification model incorporating BERT and several training strategies including dynamic negative sample selection.
Experiments on SciAD show the effectiveness of our proposed model and our score ranks 1st in SDU@AAAI-21 shared task 2: Acronym Disambiguation.
arXiv Detail & Related papers (2021-02-25T05:40:21Z) - SciDr at SDU-2020: IDEAS -- Identifying and Disambiguating Everyday
Acronyms for Scientific Domain [0.0]
We present our systems submitted for the shared tasks of Acronym Identification (AI) and Acronym Disambiguation (AD)
We mainly experiment with BERT and SciBERT.
For AD, we formulate the problem as a span prediction task, experiment with different training techniques and also leverage the use of external data.
arXiv Detail & Related papers (2021-02-17T15:24:50Z) - AT-BERT: Adversarial Training BERT for Acronym Identification Winning
Solution for SDU@AAAI-21 [5.478126869836199]
Acronym identification focuses on finding the acronyms and the phrases that have been abbreviated.
Recent breakthroughs of language models pre-trained on large corpora clearly show that unsupervised pre-training can vastly improve the performance of downstream tasks.
We present an Adversarial Training BERT method named AT-BERT, our winning solution to acronym identification task for Scientific Document Understanding (SDU) Challenge of AAAI 2021.
arXiv Detail & Related papers (2021-01-11T05:02:34Z) - Acronym Identification and Disambiguation Shared Tasks for Scientific
Document Understanding [41.63345823743157]
Acronyms are short forms of longer phrases frequently used in writing.
Every text understanding tool should be capable of recognizing acronyms in text.
To push forward research in this direction, we have organized two shared task for acronym identification and acronym disambiguation in scientific documents.
arXiv Detail & Related papers (2020-12-22T00:29:15Z) - What Does This Acronym Mean? Introducing a New Dataset for Acronym
Identification and Disambiguation [74.42107665213909]
Acronyms are the short forms of phrases that facilitate conveying lengthy sentences in documents and serve as one of the mainstays of writing.
Due to their importance, identifying acronyms and corresponding phrases (AI) and finding the correct meaning of each acronym (i.e., acronym disambiguation (AD)) are crucial for text understanding.
Despite the recent progress on this task, there are some limitations in the existing datasets which hinder further improvement.
arXiv Detail & Related papers (2020-10-28T00:12:36Z) - Interpretability Analysis for Named Entity Recognition to Understand
System Predictions and How They Can Improve [49.878051587667244]
We examine the performance of several variants of LSTM-CRF architectures for named entity recognition.
We find that context representations do contribute to system performance, but that the main factor driving high performance is learning the name tokens themselves.
We enlist human annotators to evaluate the feasibility of inferring entity types from the context alone and find that, while people are not able to infer the entity type either for the majority of the errors made by the context-only system, there is some room for improvement.
arXiv Detail & Related papers (2020-04-09T14:37:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.