NER-to-MRC: Named-Entity Recognition Completely Solving as Machine
Reading Comprehension
- URL: http://arxiv.org/abs/2305.03970v1
- Date: Sat, 6 May 2023 08:05:22 GMT
- Title: NER-to-MRC: Named-Entity Recognition Completely Solving as Machine
Reading Comprehension
- Authors: Yuxiang Zhang, Junjie Wang, Xinyu Zhu, Tetsuya Sakai, Hayato Yamana
- Abstract summary: We frame NER as a machine reading comprehension problem, called NER-to-MRC.
We transform the NER task into a form suitable for the model to solve with MRC in a efficient manner.
We achieve state-of-the-art performance without external data, up to 11.24% improvement on the WNUT-16 dataset.
- Score: 29.227500985892195
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Named-entity recognition (NER) detects texts with predefined semantic labels
and is an essential building block for natural language processing (NLP).
Notably, recent NER research focuses on utilizing massive extra data, including
pre-training corpora and incorporating search engines. However, these methods
suffer from high costs associated with data collection and pre-training, and
additional training process of the retrieved data from search engines. To
address the above challenges, we completely frame NER as a machine reading
comprehension (MRC) problem, called NER-to-MRC, by leveraging MRC with its
ability to exploit existing data efficiently. Several prior works have been
dedicated to employing MRC-based solutions for tackling the NER problem,
several challenges persist: i) the reliance on manually designed prompts; ii)
the limited MRC approaches to data reconstruction, which fails to achieve
performance on par with methods utilizing extensive additional data. Thus, our
NER-to-MRC conversion consists of two components: i) transform the NER task
into a form suitable for the model to solve with MRC in a efficient manner; ii)
apply the MRC reasoning strategy to the model. We experiment on 6 benchmark
datasets from three domains and achieve state-of-the-art performance without
external data, up to 11.24% improvement on the WNUT-16 dataset.
Related papers
- MindStar: Enhancing Math Reasoning in Pre-trained LLMs at Inference Time [51.5039731721706]
MindStar is a purely inference-based searching method for large language models.
It formulates reasoning tasks as searching problems and proposes two search ideas to identify the optimal reasoning paths.
It significantly enhances the reasoning abilities of open-source models, such as Llama-2-13B and Mistral-7B, and achieves comparable performance to GPT-3.5 and Grok-1.
arXiv Detail & Related papers (2024-05-25T15:07:33Z) - SCANNER: Knowledge-Enhanced Approach for Robust Multi-modal Named Entity Recognition of Unseen Entities [10.193908215351497]
We propose SCANNER, a model capable of effectively handling all three NER variants.
SCANNER is a two-stage structure; we extract entity candidates in the first stage and use it as a query to get knowledge.
To tackle the challenges arising from noisy annotations in NER datasets, we introduce a novel self-distillation method.
arXiv Detail & Related papers (2024-04-02T13:05:41Z) - LLM-DA: Data Augmentation via Large Language Models for Few-Shot Named
Entity Recognition [67.96794382040547]
$LLM-DA$ is a novel data augmentation technique based on large language models (LLMs) for the few-shot NER task.
Our approach involves employing 14 contextual rewriting strategies, designing entity replacements of the same type, and incorporating noise injection to enhance robustness.
arXiv Detail & Related papers (2024-02-22T14:19:56Z) - NERetrieve: Dataset for Next Generation Named Entity Recognition and
Retrieval [49.827932299460514]
We argue that capabilities provided by large language models are not the end of NER research, but rather an exciting beginning.
We present three variants of the NER task, together with a dataset to support them.
We provide a large, silver-annotated corpus of 4 million paragraphs covering 500 entity types.
arXiv Detail & Related papers (2023-10-22T12:23:00Z) - Named Entity Recognition via Machine Reading Comprehension: A Multi-Task
Learning Approach [50.12455129619845]
Named Entity Recognition (NER) aims to extract and classify entity mentions in the text into pre-defined types.
We propose to incorporate the label dependencies among entity types into a multi-task learning framework for better MRC-based NER.
arXiv Detail & Related papers (2023-09-20T03:15:05Z) - NER-MQMRC: Formulating Named Entity Recognition as Multi Question
Machine Reading Comprehension [6.3066273316843775]
We propose posing NER as a multi-question MRC task, where multiple questions are considered at the same time for a single text.
We show that our proposed architecture leads to average 2.5 times faster training and 2.3 times faster inference as compared to NER-SQMRC framework based models.
arXiv Detail & Related papers (2022-05-12T06:54:03Z) - Distantly-Supervised Named Entity Recognition with Noise-Robust Learning
and Language Model Augmented Self-Training [66.80558875393565]
We study the problem of training named entity recognition (NER) models using only distantly-labeled data.
We propose a noise-robust learning scheme comprised of a new loss function and a noisy label removal step.
Our method achieves superior performance, outperforming existing distantly-supervised NER models by significant margins.
arXiv Detail & Related papers (2021-09-10T17:19:56Z) - An Open-Source Dataset and A Multi-Task Model for Malay Named Entity
Recognition [3.511753382329252]
We build a Malay NER dataset (MYNER) comprising 28,991 sentences (over 384 thousand tokens)
An auxiliary task, boundary detection, is introduced to improve NER training in both explicit and implicit ways.
arXiv Detail & Related papers (2021-09-03T03:29:25Z) - Exploiting Neural Query Translation into Cross Lingual Information
Retrieval [49.167049709403166]
Existing CLIR systems mainly exploit statistical-based machine translation (SMT) rather than the advanced neural machine translation (NMT)
We propose a novel data augmentation method that extracts query translation pairs according to user clickthrough data.
Experimental results reveal that the proposed approach yields better retrieval quality than strong baselines.
arXiv Detail & Related papers (2020-10-26T15:28:19Z) - A Survey on Machine Reading Comprehension: Tasks, Evaluation Metrics and
Benchmark Datasets [5.54205518616467]
Machine Reading (MRC) is a challenging Natural Language Processing(NLP) research field with wide real-world applications.
A lot of MRC models have already surpassed human performance on various benchmark datasets.
This shows the need for improving existing datasets, evaluation metrics, and models to move current MRC models toward "real" understanding.
arXiv Detail & Related papers (2020-06-21T19:18:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.