NER-MQMRC: Formulating Named Entity Recognition as Multi Question
Machine Reading Comprehension
- URL: http://arxiv.org/abs/2205.05904v1
- Date: Thu, 12 May 2022 06:54:03 GMT
- Title: NER-MQMRC: Formulating Named Entity Recognition as Multi Question
Machine Reading Comprehension
- Authors: Anubhav Shrimal, Avi Jain, Kartik Mehta, Promod Yenigalla
- Abstract summary: We propose posing NER as a multi-question MRC task, where multiple questions are considered at the same time for a single text.
We show that our proposed architecture leads to average 2.5 times faster training and 2.3 times faster inference as compared to NER-SQMRC framework based models.
- Score: 6.3066273316843775
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: NER has been traditionally formulated as a sequence labeling task. However,
there has been recent trend in posing NER as a machine reading comprehension
task (Wang et al., 2020; Mengge et al., 2020), where entity name (or other
information) is considered as a question, text as the context and entity value
in text as answer snippet. These works consider MRC based on a single question
(entity) at a time. We propose posing NER as a multi-question MRC task, where
multiple questions (one question per entity) are considered at the same time
for a single text. We propose a novel BERT-based multi-question MRC (NER-MQMRC)
architecture for this formulation. NER-MQMRC architecture considers all
entities as input to BERT for learning token embeddings with self-attention and
leverages BERT-based entity representation for further improving these token
embeddings for NER task. Evaluation on three NER datasets show that our
proposed architecture leads to average 2.5 times faster training and 2.3 times
faster inference as compared to NER-SQMRC framework based models by considering
all entities together in a single pass. Further, we show that our model
performance does not degrade compared to single-question based MRC (NER-SQMRC)
(Devlin et al., 2019) leading to F1 gain of +0.41%, +0.32% and +0.27% for
AE-Pub, Ecommerce5PT and Twitter datasets respectively. We propose this
architecture primarily to solve large scale e-commerce attribute (or entity)
extraction from unstructured text of a magnitude of 50k+ attributes to be
extracted on a scalable production environment with high performance and
optimised training and inference runtimes.
Related papers
- Advancing Grounded Multimodal Named Entity Recognition via LLM-Based Reformulation and Box-Based Segmentation [46.9782192992495]
Grounded Multimodal Named Entity Recognition (GMNER) task aims to identify named entities, entity types and their corresponding visual regions.
We propose RiVEG, a unified framework that reformulates GMNER into a joint MNER-VE-VG task by leveraging large language models.
arXiv Detail & Related papers (2024-06-11T13:52:29Z) - NERetrieve: Dataset for Next Generation Named Entity Recognition and
Retrieval [49.827932299460514]
We argue that capabilities provided by large language models are not the end of NER research, but rather an exciting beginning.
We present three variants of the NER task, together with a dataset to support them.
We provide a large, silver-annotated corpus of 4 million paragraphs covering 500 entity types.
arXiv Detail & Related papers (2023-10-22T12:23:00Z) - Named Entity Recognition via Machine Reading Comprehension: A Multi-Task
Learning Approach [50.12455129619845]
Named Entity Recognition (NER) aims to extract and classify entity mentions in the text into pre-defined types.
We propose to incorporate the label dependencies among entity types into a multi-task learning framework for better MRC-based NER.
arXiv Detail & Related papers (2023-09-20T03:15:05Z) - NER-to-MRC: Named-Entity Recognition Completely Solving as Machine
Reading Comprehension [29.227500985892195]
We frame NER as a machine reading comprehension problem, called NER-to-MRC.
We transform the NER task into a form suitable for the model to solve with MRC in a efficient manner.
We achieve state-of-the-art performance without external data, up to 11.24% improvement on the WNUT-16 dataset.
arXiv Detail & Related papers (2023-05-06T08:05:22Z) - MNER-QG: An End-to-End MRC framework for Multimodal Named Entity
Recognition with Query Grounding [21.49274082010887]
Multimodal named entity recognition (MNER) is a critical step in information extraction.
We propose a novel end-to-end framework named MNER-QG that can simultaneously perform MRC-based MRC-based named entity recognition and query grounding.
arXiv Detail & Related papers (2022-11-27T06:10:03Z) - MultiCoNER: A Large-scale Multilingual dataset for Complex Named Entity
Recognition [15.805414696789796]
We present MultiCoNER, a large multilingual dataset for Named Entity Recognition that covers 3 domains (Wiki sentences, questions, and search queries) across 11 languages.
This dataset is designed to represent contemporary challenges in NER, including low-context scenarios.
arXiv Detail & Related papers (2022-08-30T20:45:54Z) - Pack Together: Entity and Relation Extraction with Levitated Marker [61.232174424421025]
We propose a novel span representation approach, named Packed Levitated Markers, to consider the dependencies between the spans (pairs) by strategically packing the markers in the encoder.
Our experiments show that our model with packed levitated markers outperforms the sequence labeling model by 0.4%-1.9% F1 on three flat NER tasks, and beats the token concat model on six NER benchmarks.
arXiv Detail & Related papers (2021-09-13T15:38:13Z) - Data Augmentation for Abstractive Query-Focused Multi-Document
Summarization [129.96147867496205]
We present two QMDS training datasets, which we construct using two data augmentation methods.
These two datasets have complementary properties, i.e., QMDSCNN has real summaries but queries are simulated, while QMDSIR has real queries but simulated summaries.
We build end-to-end neural network models on the combined datasets that yield new state-of-the-art transfer results on DUC datasets.
arXiv Detail & Related papers (2021-03-02T16:57:01Z) - Structured Multimodal Attentions for TextVQA [57.71060302874151]
We propose an end-to-end structured multimodal attention (SMA) neural network to mainly solve the first two issues above.
SMA first uses a structural graph representation to encode the object-object, object-text and text-text relationships appearing in the image, and then designs a multimodal graph attention network to reason over it.
Our proposed model outperforms the SoTA models on TextVQA dataset and two tasks of ST-VQA dataset among all models except pre-training based TAP.
arXiv Detail & Related papers (2020-06-01T07:07:36Z) - One Model to Recognize Them All: Marginal Distillation from NER Models
with Different Tag Sets [30.445201832698192]
Named entity recognition (NER) is a fundamental component in the modern language understanding pipeline.
This paper presents a marginal distillation (MARDI) approach for training a unified NER model from resources with disjoint or heterogeneous tag sets.
arXiv Detail & Related papers (2020-04-10T17:36:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.