T-NER: An All-Round Python Library for Transformer-based Named Entity
Recognition
- URL: http://arxiv.org/abs/2209.12616v1
- Date: Fri, 9 Sep 2022 15:00:38 GMT
- Title: T-NER: An All-Round Python Library for Transformer-based Named Entity
Recognition
- Authors: Asahi Ushio, Jose Camacho-Collados
- Abstract summary: T-NER is a Python library for NER LM finetuning.
We show the potential of the library by compiling nine public NER datasets into a unified format.
To facilitate future research, we also release all our LM checkpoints via the Hugging Face model hub.
- Score: 9.928025283928282
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Language model (LM) pretraining has led to consistent improvements in many
NLP downstream tasks, including named entity recognition (NER). In this paper,
we present T-NER (Transformer-based Named Entity Recognition), a Python library
for NER LM finetuning. In addition to its practical utility, T-NER facilitates
the study and investigation of the cross-domain and cross-lingual
generalization ability of LMs finetuned on NER. Our library also provides a web
app where users can get model predictions interactively for arbitrary text,
which facilitates qualitative model evaluation for non-expert programmers. We
show the potential of the library by compiling nine public NER datasets into a
unified format and evaluating the cross-domain and cross-lingual performance
across the datasets. The results from our initial experiments show that
in-domain performance is generally competitive across datasets. However,
cross-domain generalization is challenging even with a large pretrained LM,
which has nevertheless capacity to learn domain-specific features if fine-tuned
on a combined dataset. To facilitate future research, we also release all our
LM checkpoints via the Hugging Face model hub.
Related papers
- Improving Few-Shot Cross-Domain Named Entity Recognition by Instruction Tuning a Word-Embedding based Retrieval Augmented Large Language Model [0.0]
Few-Shot Cross-Domain NER is a process of leveraging knowledge from data-rich source domains to perform entity recognition on data scarce target domains.
We propose IF-WRANER, a retrieval augmented large language model for Named Entity Recognition.
arXiv Detail & Related papers (2024-11-01T08:57:29Z) - llmNER: (Zero|Few)-Shot Named Entity Recognition, Exploiting the Power of Large Language Models [1.1196013962698619]
This paper presents llmNER, a Python library for implementing zero-shot and few-shot NER with large language models (LLMs)
llmNER can compose prompts, query the model, and parse the completion returned by the LLM.
We validated our software on two NER tasks to show the library's flexibility.
arXiv Detail & Related papers (2024-06-06T22:01:59Z) - Learning to Rank Context for Named Entity Recognition Using a Synthetic Dataset [6.633914491587503]
We propose to generate a synthetic context retrieval training dataset using Alpaca.
Using this dataset, we train a neural context retriever based on a BERT model that is able to find relevant context for NER.
We show that our method outperforms several retrieval baselines for the NER task on an English literary dataset composed of the first chapter of 40 books.
arXiv Detail & Related papers (2023-10-16T06:53:12Z) - A Confidence-based Partial Label Learning Model for Crowd-Annotated
Named Entity Recognition [74.79785063365289]
Existing models for named entity recognition (NER) are mainly based on large-scale labeled datasets.
We propose a Confidence-based Partial Label Learning (CPLL) method to integrate the prior confidence (given by annotators) and posterior confidences (learned by models) for crowd-annotated NER.
arXiv Detail & Related papers (2023-05-21T15:31:23Z) - One Model for All Domains: Collaborative Domain-Prefix Tuning for
Cross-Domain NER [92.79085995361098]
Cross-domain NER is a challenging task to address the low-resource problem in practical scenarios.
Previous solutions mainly obtain a NER model by pre-trained language models (PLMs) with data from a rich-resource domain and adapt it to the target domain.
We introduce Collaborative Domain-Prefix Tuning for cross-domain NER based on text-to-text generative PLMs.
arXiv Detail & Related papers (2023-01-25T05:16:43Z) - Domain-Specific NER via Retrieving Correlated Samples [37.98414661072985]
In this paper, we suggest enhancing NER models with correlated samples.
To explicitly simulate the human reasoning process, we perform a training-free entity type calibrating by majority voting.
Empirical results on datasets of the above two domains show the efficacy of our methods.
arXiv Detail & Related papers (2022-08-27T12:25:24Z) - Non-Parametric Unsupervised Domain Adaptation for Neural Machine
Translation [61.27321597981737]
$k$NN-MT has shown the promising capability of directly incorporating the pre-trained neural machine translation (NMT) model with domain-specific token-level $k$-nearest-neighbor retrieval.
We propose a novel framework that directly uses in-domain monolingual sentences in the target language to construct an effective datastore for $k$-nearest-neighbor retrieval.
arXiv Detail & Related papers (2021-09-14T11:50:01Z) - An Open-Source Dataset and A Multi-Task Model for Malay Named Entity
Recognition [3.511753382329252]
We build a Malay NER dataset (MYNER) comprising 28,991 sentences (over 384 thousand tokens)
An auxiliary task, boundary detection, is introduced to improve NER training in both explicit and implicit ways.
arXiv Detail & Related papers (2021-09-03T03:29:25Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - Local Additivity Based Data Augmentation for Semi-supervised NER [59.90773003737093]
Named Entity Recognition (NER) is one of the first stages in deep language understanding.
Current NER models heavily rely on human-annotated data.
We propose a Local Additivity based Data Augmentation (LADA) method for semi-supervised NER.
arXiv Detail & Related papers (2020-10-04T20:46:26Z) - Zero-Resource Cross-Domain Named Entity Recognition [68.83177074227598]
Existing models for cross-domain named entity recognition rely on numerous unlabeled corpus or labeled NER training data in target domains.
We propose a cross-domain NER model that does not use any external resources.
arXiv Detail & Related papers (2020-02-14T09:04:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.