$k$NN-NER: Named Entity Recognition with Nearest Neighbor Search
- URL: http://arxiv.org/abs/2203.17103v1
- Date: Thu, 31 Mar 2022 15:21:43 GMT
- Title: $k$NN-NER: Named Entity Recognition with Nearest Neighbor Search
- Authors: Shuhe Wang, Xiaoya Li, Yuxian Meng, Tianwei Zhang, Rongbin Ouyang,
Jiwei Li, Guoyin Wang
- Abstract summary: $k$ nearest neighbor NER ($k$NN-NER) framework augments the distribution of entity labels by assigning $k$ nearest neighbors retrieved from the training set.
$k$NN-NER requires no additional operation during the training phase, and by interpolating $k$ nearest neighbors search into the vanilla NER model, $k$NN-NER consistently outperforms its vanilla counterparts.
- Score: 47.901071142524906
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inspired by recent advances in retrieval augmented methods in
NLP~\citep{khandelwal2019generalization,khandelwal2020nearest,meng2021gnn}, in
this paper, we introduce a $k$ nearest neighbor NER ($k$NN-NER) framework,
which augments the distribution of entity labels by assigning $k$ nearest
neighbors retrieved from the training set. This strategy makes the model more
capable of handling long-tail cases, along with better few-shot learning
abilities. $k$NN-NER requires no additional operation during the training
phase, and by interpolating $k$ nearest neighbors search into the vanilla NER
model, $k$NN-NER consistently outperforms its vanilla counterparts: we achieve
a new state-of-the-art F1-score of 72.03 (+1.25) on the Chinese Weibo dataset
and improved results on a variety of widely used NER benchmarks. Additionally,
we show that $k$NN-NER can achieve comparable results to the vanilla NER model
with 40\% less amount of training data. Code available at
\url{https://github.com/ShannonAI/KNN-NER}.
Related papers
- INK: Injecting kNN Knowledge in Nearest Neighbor Machine Translation [57.952478914459164]
kNN-MT has provided an effective paradigm to smooth the prediction based on neighbor representations during inference.
We propose an effective training framework INK to directly smooth the representation space via adjusting representations of kNN neighbors with a small number of new parameters.
Experiments on four benchmark datasets show that method achieves average gains of 1.99 COMET and 1.0 BLEU, outperforming the state-of-the-art kNN-MT system with 0.02x memory space and 1.9x inference speedup.
arXiv Detail & Related papers (2023-06-10T08:39:16Z) - PromptNER: Prompting For Named Entity Recognition [27.501500279749475]
We introduce PromptNER, a new state-of-the-art algorithm for few-Shot and cross-domain NER.
PromptNER achieves a 4% (absolute) improvement in F1 score on the ConLL dataset, a 9% (absolute) improvement on the GENIA dataset, and a 4% (absolute) improvement on the FewNERD dataset.
arXiv Detail & Related papers (2023-05-24T07:38:24Z) - An Embarrassingly Easy but Strong Baseline for Nested Named Entity
Recognition [55.080101447586635]
We propose using Conal Neural Network (CNN) to model spatial relations in the score matrix.
Our model surpasses several recently proposed methods with the same pre-trained encoders.
arXiv Detail & Related papers (2022-08-09T04:33:46Z) - NERDA-Con: Extending NER models for Continual Learning -- Integrating
Distinct Tasks and Updating Distribution Shifts [0.0]
We propose NERDA-Con, a pipeline for training NERs with Large Language Models (LLMs) bases.
As we believe our work has implications to be utilized in the pipeline of continual learning and NER, we open-source our code as well as provide the fine-tuning library of the same name NERDA-Con.
arXiv Detail & Related papers (2022-06-28T03:22:55Z) - Nearest Neighbor Zero-Shot Inference [68.56747574377215]
kNN-Prompt is a technique to use k-nearest neighbor (kNN) retrieval augmentation for zero-shot inference with language models (LMs)
fuzzy verbalizers leverage the sparse kNN distribution for downstream tasks by automatically associating each classification label with a set of natural language tokens.
Experiments show that kNN-Prompt is effective for domain adaptation with no further training, and that the benefits of retrieval increase with the size of the model used for kNN retrieval.
arXiv Detail & Related papers (2022-05-27T07:00:59Z) - Unified Named Entity Recognition as Word-Word Relation Classification [25.801945832005504]
We present a novel alternative by modeling the unified NER as word-word relation classification, namely W2NER.
The architecture resolves the kernel bottleneck of unified NER by effectively modeling the neighboring relations between entity words.
Based on the W2NER scheme we develop a neural framework, in which the unified NER is modeled as a 2D grid of word pairs.
arXiv Detail & Related papers (2021-12-19T06:11:07Z) - Under-bagging Nearest Neighbors for Imbalanced Classification [63.026765294759876]
We propose an ensemble learning algorithm called textitunder-bagging $k$-NN (textitunder-bagging $k$-NN) for imbalanced classification problems.
arXiv Detail & Related papers (2021-09-01T14:10:38Z) - Adaptive Nearest Neighbor Machine Translation [60.97183408140499]
kNN-MT combines pre-trained neural machine translation with token-level k-nearest-neighbor retrieval.
Traditional kNN algorithm simply retrieves a same number of nearest neighbors for each target token.
We propose Adaptive kNN-MT to dynamically determine the number of k for each target token.
arXiv Detail & Related papers (2021-05-27T09:27:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.