Beheshti-NER: Persian Named Entity Recognition Using BERT
- URL: http://arxiv.org/abs/2003.08875v1
- Date: Thu, 19 Mar 2020 15:55:21 GMT
- Title: Beheshti-NER: Persian Named Entity Recognition Using BERT
- Authors: Ehsan Taher, Seyed Abbas Hoseini, and Mehrnoush Shamsfard
- Abstract summary: In this paper, we use the pre-trained deep bidirectional network, BERT, to make a model for named entity recognition in Persian.
Our results are 83.5 and 88.4 f1 CONLL score respectively in phrase and word level evaluation.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Named entity recognition is a natural language processing task to recognize
and extract spans of text associated with named entities and classify them in
semantic Categories.
Google BERT is a deep bidirectional language model, pre-trained on large
corpora that can be fine-tuned to solve many NLP tasks such as question
answering, named entity recognition, part of speech tagging and etc. In this
paper, we use the pre-trained deep bidirectional network, BERT, to make a model
for named entity recognition in Persian.
We also compare the results of our model with the previous state of the art
results achieved on Persian NER. Our evaluation metric is CONLL 2003 score in
two levels of word and phrase. This model achieved second place in NSURL-2019
task 7 competition which associated with NER for the Persian language. our
results in this competition are 83.5 and 88.4 f1 CONLL score respectively in
phrase and word level evaluation.
Related papers
- A Generative Approach for Wikipedia-Scale Visual Entity Recognition [56.55633052479446]
We address the task of mapping a given query image to one of the 6 million existing entities in Wikipedia.
We introduce a novel Generative Entity Recognition framework, which learns to auto-regressively decode a semantic and discriminative code'' identifying the target entity.
arXiv Detail & Related papers (2024-03-04T13:47:30Z) - Multicultural Name Recognition For Previously Unseen Names [65.268245109828]
This paper attempts to improve recognition of person names, a diverse category that can grow any time someone is born or changes their name.
I look at names from 103 countries to compare how well the model performs on names from different cultures.
I find that a model with combined character and word input outperforms word-only models and may improve on accuracy compared to classical NER models.
arXiv Detail & Related papers (2024-01-23T17:58:38Z) - IXA/Cogcomp at SemEval-2023 Task 2: Context-enriched Multilingual Named
Entity Recognition using Knowledge Bases [53.054598423181844]
We present a novel NER cascade approach comprising three steps.
We empirically demonstrate the significance of external knowledge bases in accurately classifying fine-grained and emerging entities.
Our system exhibits robust performance in the MultiCoNER2 shared task, even in the low-resource language setting.
arXiv Detail & Related papers (2023-04-20T20:30:34Z) - Disambiguation of Company names via Deep Recurrent Networks [101.90357454833845]
We propose a Siamese LSTM Network approach to extract -- via supervised learning -- an embedding of company name strings.
We analyse how an Active Learning approach to prioritise the samples to be labelled leads to a more efficient overall learning pipeline.
arXiv Detail & Related papers (2023-03-07T15:07:57Z) - ANEC: An Amharic Named Entity Corpus and Transformer Based Recognizer [0.0]
We present an Amharic named entity recognition system based on bidirectional long short-term memory with a conditional random fields layer.
Our named entity recognition system achieves an F_1 score of 93%, which is the new state-of-the-art result for Amharic named entity recognition.
arXiv Detail & Related papers (2022-07-02T09:50:37Z) - Label Semantics for Few Shot Named Entity Recognition [68.01364012546402]
We study the problem of few shot learning for named entity recognition.
We leverage the semantic information in the names of the labels as a way of giving the model additional signal and enriched priors.
Our model learns to match the representations of named entities computed by the first encoder with label representations computed by the second encoder.
arXiv Detail & Related papers (2022-03-16T23:21:05Z) - WCL-BBCD: A Contrastive Learning and Knowledge Graph Approach to Named
Entity Recognition [15.446770390648874]
We propose a novel named entity recognition model WCL-BBCD (Word Contrastive Learning with BERT-BiLSTM-CRF-DBpedia)
The model first trains the sentence pairs in the text, calculate similarity between words in sentence pairs by cosine similarity, and fine-tunes the BERT model used for the named entity recognition task through the similarity.
Finally, the recognition results are corrected in combination with prior knowledge such as knowledge graphs, so as to alleviate the recognition caused by word abbreviations low-rate problem.
arXiv Detail & Related papers (2022-03-14T08:29:58Z) - A Sequence-to-Set Network for Nested Named Entity Recognition [38.05786148160635]
We propose a novel sequence-to-set neural network for nested NER.
We use a non-autoregressive decoder to predict the final set of entities in one pass.
Experimental results show that our proposed model achieves state-of-the-art on three nested NER corpora.
arXiv Detail & Related papers (2021-05-19T03:10:04Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - Named Entity Recognition as Dependency Parsing [16.544333689188246]
We use graph-based dependency parsing to provide our model a global view on the input via a biaffine model.
We show that the model works well for both nested and flat NER through evaluation on 8 corpora and achieving SoTA performance on all of them, with accuracy gains of up to 2.2 percentage points.
arXiv Detail & Related papers (2020-05-14T17:11:41Z) - Integrating Boundary Assembling into a DNN Framework for Named Entity
Recognition in Chinese Social Media Text [3.7239227834407735]
Chinese word boundaries are also entity boundaries, so named entity recognition for Chinese text can benefit from word boundary detection.
In this paper, we integrate a boundary assembling method with the state-of-the-art deep neural network model, and incorporate the updated word boundary information into a conditional random field model for named entity recognition.
Our method shows a 2% absolute improvement over previous state-of-the-art results.
arXiv Detail & Related papers (2020-02-27T04:29:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.