PromptNER: Prompt Locating and Typing for Named Entity Recognition
- URL: http://arxiv.org/abs/2305.17104v1
- Date: Fri, 26 May 2023 17:16:11 GMT
- Title: PromptNER: Prompt Locating and Typing for Named Entity Recognition
- Authors: Yongliang Shen, Zeqi Tan, Shuhui Wu, Wenqi Zhang, Rongsheng Zhang,
Yadong Xi, Weiming Lu, Yueting Zhuang
- Abstract summary: In this paper, we design a dual-slot multi-prompt template with the position slot and type slot to prompt locating and typing respectively.
Multiple prompts can be input to the model simultaneously, and then the model extracts all entities by parallel predictions on the slots.
Experimental results show that the proposed model achieves a significant performance improvement, especially in the cross-domain few-shot setting.
- Score: 39.81221703760443
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Prompt learning is a new paradigm for utilizing pre-trained language models
and has achieved great success in many tasks. To adopt prompt learning in the
NER task, two kinds of methods have been explored from a pair of symmetric
perspectives, populating the template by enumerating spans to predict their
entity types or constructing type-specific prompts to locate entities. However,
these methods not only require a multi-round prompting manner with a high time
overhead and computational cost, but also require elaborate prompt templates,
that are difficult to apply in practical scenarios. In this paper, we unify
entity locating and entity typing into prompt learning, and design a dual-slot
multi-prompt template with the position slot and type slot to prompt locating
and typing respectively. Multiple prompts can be input to the model
simultaneously, and then the model extracts all entities by parallel
predictions on the slots. To assign labels for the slots during training, we
design a dynamic template filling mechanism that uses the extended bipartite
graph matching between prompts and the ground-truth entities. We conduct
experiments in various settings, including resource-rich flat and nested NER
datasets and low-resource in-domain and cross-domain datasets. Experimental
results show that the proposed model achieves a significant performance
improvement, especially in the cross-domain few-shot setting, which outperforms
the state-of-the-art model by +7.7% on average.
Related papers
- Automated Few-shot Classification with Instruction-Finetuned Language
Models [76.69064714392165]
We show that AuT-Few outperforms state-of-the-art few-shot learning methods.
We also show that AuT-Few is the best ranking method across datasets on the RAFT few-shot benchmark.
arXiv Detail & Related papers (2023-05-21T21:50:27Z) - PromptNER: A Prompting Method for Few-shot Named Entity Recognition via
k Nearest Neighbor Search [56.81939214465558]
We propose PromptNER: a novel prompting method for few-shot NER via k nearest neighbor search.
We use prompts that contains entity category information to construct label prototypes, which enables our model to fine-tune with only the support set.
Our approach achieves excellent transfer learning ability, and extensive experiments on the Few-NERD and CrossNER datasets demonstrate that our model achieves superior performance over state-of-the-art methods.
arXiv Detail & Related papers (2023-05-20T15:47:59Z) - STPrompt: Semantic-guided and Task-driven prompts for Effective Few-shot
Classification [5.6205035780719275]
We propose the STPrompt -Semantic-guided and Task-driven Prompt model.
The proposed model achieves the state-of-the-art performance in five different datasets of few-shot text classification tasks.
arXiv Detail & Related papers (2022-10-29T04:42:30Z) - Few-Shot Fine-Grained Entity Typing with Automatic Label Interpretation
and Instance Generation [36.541309948222306]
We study the problem of few-shot Fine-grained Entity Typing (FET), where only a few annotated entity mentions with contexts are given for each entity type.
We propose a novel framework for few-shot FET consisting of two modules: (1) an entity type label interpretation module automatically learns to relate type labels to the vocabulary by jointly leveraging few-shot instances and the label hierarchy, and (2) a type-based contextualized instance generator produces new instances based on given instances to enlarge the training set for better generalization.
arXiv Detail & Related papers (2022-06-28T04:05:40Z) - An Information-theoretic Approach to Prompt Engineering Without Ground
Truth Labels [55.06990011183662]
We introduce a new method for selecting prompt templates textitwithout labeled examples and textitwithout direct access to the model.
Across 8 datasets representing 7 distinct NLP tasks, we show that when a template has high mutual information, it also has high accuracy on the task.
arXiv Detail & Related papers (2022-03-21T21:51:43Z) - InstructionNER: A Multi-Task Instruction-Based Generative Framework for
Few-shot NER [31.32381919473188]
We propose a multi-task instruction-based generative framework, named InstructionNER, for low-resource named entity recognition.
Specifically, we reformulate the NER task as a generation problem, which enriches source sentences with task-specific instructions and answer options, then inferences the entities and types in natural language.
Experimental results show that our method consistently outperforms other baselines on five datasets in few-shot settings.
arXiv Detail & Related papers (2022-03-08T07:56:36Z) - X2Parser: Cross-Lingual and Cross-Domain Framework for Task-Oriented
Compositional Semantic Parsing [51.81533991497547]
Task-oriented compositional semantic parsing (TCSP) handles complex nested user queries.
We present X2 compared a transferable Cross-lingual and Cross-domain for TCSP.
We propose to predict flattened intents and slots representations separately and cast both prediction tasks into sequence labeling problems.
arXiv Detail & Related papers (2021-06-07T16:40:05Z) - Interpretable Entity Representations through Large-Scale Typing [61.4277527871572]
We present an approach to creating entity representations that are human readable and achieve high performance out of the box.
Our representations are vectors whose values correspond to posterior probabilities over fine-grained entity types.
We show that it is possible to reduce the size of our type set in a learning-based way for particular domains.
arXiv Detail & Related papers (2020-04-30T23:58:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.