Multicultural Name Recognition For Previously Unseen Names
- URL: http://arxiv.org/abs/2401.12941v1
- Date: Tue, 23 Jan 2024 17:58:38 GMT
- Title: Multicultural Name Recognition For Previously Unseen Names
- Authors: Alexandra Loessberg-Zahl
- Abstract summary: This paper attempts to improve recognition of person names, a diverse category that can grow any time someone is born or changes their name.
I look at names from 103 countries to compare how well the model performs on names from different cultures.
I find that a model with combined character and word input outperforms word-only models and may improve on accuracy compared to classical NER models.
- Score: 65.268245109828
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: State of the art Named Entity Recognition (NER) models have achieved an
impressive ability to extract common phrases from text that belong to labels
such as location, organization, time, and person. However, typical NER systems
that rely on having seen a specific entity in their training data in order to
label an entity perform poorly on rare or unseen entities ta in order to label
an entity perform poorly on rare or unseen entities (Derczynski et al., 2017).
This paper attempts to improve recognition of person names, a diverse category
that can grow any time someone is born or changes their name. In order for
downstream tasks to not exhibit bias based on cultural background, a model
should perform well on names from a variety of backgrounds. In this paper I
experiment with the training data and input structure of an English Bi-LSTM
name recognition model. I look at names from 103 countries to compare how well
the model performs on names from different cultures, specifically in the
context of a downstream task where extracted names will be matched to
information on file. I find that a model with combined character and word input
outperforms word-only models and may improve on accuracy compared to classical
NER models that are not geared toward identifying unseen entity values.
Related papers
- Large-Scale Label Interpretation Learning for Few-Shot Named Entity Recognition [5.262708162539423]
Few-shot named entity recognition (NER) detects named entities within text using only a few examples.
One promising line of research is to leverage natural language descriptions of each entity type.
In this paper, we explore the impact of a strong semantic prior to interpret verbalizations of new entity types.
arXiv Detail & Related papers (2024-03-21T08:22:44Z) - Using LSTM and GRU With a New Dataset for Named Entity Recognition in
the Arabic Language [0.0]
We use the BIOES format to tag the word, which allows us to handle the nested name entity.
This work proposes long short term memory (LSTM) units and Gated Recurrent Units (GRU) for building the named entity recognition model in the Arabic language.
arXiv Detail & Related papers (2023-04-06T22:14:02Z) - Disambiguation of Company names via Deep Recurrent Networks [101.90357454833845]
We propose a Siamese LSTM Network approach to extract -- via supervised learning -- an embedding of company name strings.
We analyse how an Active Learning approach to prioritise the samples to be labelled leads to a more efficient overall learning pipeline.
arXiv Detail & Related papers (2023-03-07T15:07:57Z) - Label Semantics for Few Shot Named Entity Recognition [68.01364012546402]
We study the problem of few shot learning for named entity recognition.
We leverage the semantic information in the names of the labels as a way of giving the model additional signal and enriched priors.
Our model learns to match the representations of named entities computed by the first encoder with label representations computed by the second encoder.
arXiv Detail & Related papers (2022-03-16T23:21:05Z) - A Realistic Study of Auto-regressive Language Models for Named Entity
Typing and Recognition [7.345578385749421]
We study pre-trained language models for named entity recognition in a meta-learning setup.
First, we test named entity typing (NET) in a zero-shot transfer scenario. Then, we perform NER by giving few examples at inference.
We propose a method to select seen and rare / unseen names when having access only to the pre-trained model and report results on these groups.
arXiv Detail & Related papers (2021-08-26T15:29:00Z) - More Than Words: Collocation Tokenization for Latent Dirichlet
Allocation Models [71.42030830910227]
We propose a new metric for measuring the clustering quality in settings where the models differ.
We show that topics trained with merged tokens result in topic keys that are clearer, more coherent, and more effective at distinguishing topics than those unmerged models.
arXiv Detail & Related papers (2021-08-24T14:08:19Z) - Text Classification Using Label Names Only: A Language Model
Self-Training Approach [80.63885282358204]
Current text classification methods typically require a good number of human-labeled documents as training data.
We show that our model achieves around 90% accuracy on four benchmark datasets including topic and sentiment classification.
arXiv Detail & Related papers (2020-10-14T17:06:41Z) - Interpretability Analysis for Named Entity Recognition to Understand
System Predictions and How They Can Improve [49.878051587667244]
We examine the performance of several variants of LSTM-CRF architectures for named entity recognition.
We find that context representations do contribute to system performance, but that the main factor driving high performance is learning the name tokens themselves.
We enlist human annotators to evaluate the feasibility of inferring entity types from the context alone and find that, while people are not able to infer the entity type either for the majority of the errors made by the context-only system, there is some room for improvement.
arXiv Detail & Related papers (2020-04-09T14:37:12Z) - Beheshti-NER: Persian Named Entity Recognition Using BERT [0.0]
In this paper, we use the pre-trained deep bidirectional network, BERT, to make a model for named entity recognition in Persian.
Our results are 83.5 and 88.4 f1 CONLL score respectively in phrase and word level evaluation.
arXiv Detail & Related papers (2020-03-19T15:55:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.