PromptNER: A Prompting Method for Few-shot Named Entity Recognition via
k Nearest Neighbor Search
- URL: http://arxiv.org/abs/2305.12217v1
- Date: Sat, 20 May 2023 15:47:59 GMT
- Title: PromptNER: A Prompting Method for Few-shot Named Entity Recognition via
k Nearest Neighbor Search
- Authors: Mozhi Zhang, Hang Yan, Yaqian Zhou, Xipeng Qiu
- Abstract summary: We propose PromptNER: a novel prompting method for few-shot NER via k nearest neighbor search.
We use prompts that contains entity category information to construct label prototypes, which enables our model to fine-tune with only the support set.
Our approach achieves excellent transfer learning ability, and extensive experiments on the Few-NERD and CrossNER datasets demonstrate that our model achieves superior performance over state-of-the-art methods.
- Score: 56.81939214465558
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot Named Entity Recognition (NER) is a task aiming to identify named
entities via limited annotated samples. Recently, prototypical networks have
shown promising performance in few-shot NER. Most of prototypical networks will
utilize the entities from the support set to construct label prototypes and use
the query set to compute span-level similarities and optimize these label
prototype representations. However, these methods are usually unsuitable for
fine-tuning in the target domain, where only the support set is available. In
this paper, we propose PromptNER: a novel prompting method for few-shot NER via
k nearest neighbor search. We use prompts that contains entity category
information to construct label prototypes, which enables our model to fine-tune
with only the support set. Our approach achieves excellent transfer learning
ability, and extensive experiments on the Few-NERD and CrossNER datasets
demonstrate that our model achieves superior performance over state-of-the-art
methods.
Related papers
- ToNER: Type-oriented Named Entity Recognition with Generative Language Model [14.11486479935094]
We propose a novel NER framework, namely ToNER based on a generative model.
In ToNER, a type matching model is proposed at first to identify the entity types most likely to appear in the sentence.
We append a multiple binary classification task to fine-tune the generative model's encoder, so as to generate the refined representation of the input sentence.
arXiv Detail & Related papers (2024-04-14T05:13:37Z) - NERetrieve: Dataset for Next Generation Named Entity Recognition and
Retrieval [49.827932299460514]
We argue that capabilities provided by large language models are not the end of NER research, but rather an exciting beginning.
We present three variants of the NER task, together with a dataset to support them.
We provide a large, silver-annotated corpus of 4 million paragraphs covering 500 entity types.
arXiv Detail & Related papers (2023-10-22T12:23:00Z) - MProto: Multi-Prototype Network with Denoised Optimal Transport for
Distantly Supervised Named Entity Recognition [75.87566793111066]
We propose a noise-robust prototype network named MProto for the DS-NER task.
MProto represents each entity type with multiple prototypes to characterize the intra-class variance.
To mitigate the noise from incomplete labeling, we propose a novel denoised optimal transport (DOT) algorithm.
arXiv Detail & Related papers (2023-10-12T13:02:34Z) - Named Entity Recognition via Machine Reading Comprehension: A Multi-Task
Learning Approach [50.12455129619845]
Named Entity Recognition (NER) aims to extract and classify entity mentions in the text into pre-defined types.
We propose to incorporate the label dependencies among entity types into a multi-task learning framework for better MRC-based NER.
arXiv Detail & Related papers (2023-09-20T03:15:05Z) - Type-Aware Decomposed Framework for Few-Shot Named Entity Recognition [12.444196348710728]
We propose a novel Type-Aware Decomposed framework, namely TadNER, to solve these problems.
We first present a type-aware span filtering strategy to filter out false spans by removing those semantically far away from type names.
We then present a type-aware contrastive learning strategy to construct more accurate and stable prototypes.
arXiv Detail & Related papers (2023-02-13T14:36:21Z) - Few-shot Named Entity Recognition with Entity-level Prototypical Network
Enhanced by Dispersedly Distributed Prototypes [11.759479318478915]
Few-shot named entity recognition (NER) enables us to build a NER system for a new domain using very few labeled examples.
Existing networks for this task suffer from roughly estimated label dependency and closely distributed prototypes.
EP-Net builds entity-level prototypes and considers text spans to be candidate entities, so it no longer requires the label dependency.
arXiv Detail & Related papers (2022-08-17T02:09:15Z) - Decomposed Meta-Learning for Few-Shot Named Entity Recognition [32.515795881027074]
Few-shot named entity recognition (NER) systems aim at recognizing novel-class named entities based on only a few labeled examples.
We present a meta-learning approach which tackles few-shot span detection and few-shot entity typing using meta-learning.
arXiv Detail & Related papers (2022-04-12T12:46:23Z) - Part-aware Prototype Network for Few-shot Semantic Segmentation [50.581647306020095]
We propose a novel few-shot semantic segmentation framework based on the prototype representation.
Our key idea is to decompose the holistic class representation into a set of part-aware prototypes.
We develop a novel graph neural network model to generate and enhance the proposed part-aware prototypes.
arXiv Detail & Related papers (2020-07-13T11:03:09Z) - Zero-Resource Cross-Domain Named Entity Recognition [68.83177074227598]
Existing models for cross-domain named entity recognition rely on numerous unlabeled corpus or labeled NER training data in target domains.
We propose a cross-domain NER model that does not use any external resources.
arXiv Detail & Related papers (2020-02-14T09:04:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.