Domain-Specific NER via Retrieving Correlated Samples
- URL: http://arxiv.org/abs/2208.12995v1
- Date: Sat, 27 Aug 2022 12:25:24 GMT
- Title: Domain-Specific NER via Retrieving Correlated Samples
- Authors: Xin Zhang, Yong Jiang, Xiaobin Wang, Xuming Hu, Yueheng Sun, Pengjun
Xie, Meishan Zhang
- Abstract summary: In this paper, we suggest enhancing NER models with correlated samples.
To explicitly simulate the human reasoning process, we perform a training-free entity type calibrating by majority voting.
Empirical results on datasets of the above two domains show the efficacy of our methods.
- Score: 37.98414661072985
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Successful Machine Learning based Named Entity Recognition models could fail
on texts from some special domains, for instance, Chinese addresses and
e-commerce titles, where requires adequate background knowledge. Such texts are
also difficult for human annotators. In fact, we can obtain some potentially
helpful information from correlated texts, which have some common entities, to
help the text understanding. Then, one can easily reason out the correct answer
by referencing correlated samples. In this paper, we suggest enhancing NER
models with correlated samples. We draw correlated samples by the sparse BM25
retriever from large-scale in-domain unlabeled data. To explicitly simulate the
human reasoning process, we perform a training-free entity type calibrating by
majority voting. To capture correlation features in the training stage, we
suggest to model correlated samples by the transformer-based multi-instance
cross-encoder. Empirical results on datasets of the above two domains show the
efficacy of our methods.
Related papers
- Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - Fighting Against the Repetitive Training and Sample Dependency Problem in Few-shot Named Entity Recognition [29.969268584918105]
Few-shot named entity recognition (NER) systems recognize entities using a few labeled training examples.
Current span detectors rely on extensive manual labeling to guide training.
We introduce a steppingstone span detector that is pre-trained on open-domain Wikipedia data.
We leverage a large language model (LLM) to set reliable entity-type referents, eliminating reliance on few-shot samples of each type.
arXiv Detail & Related papers (2024-06-08T12:36:30Z) - Explaining Cross-Domain Recognition with Interpretable Deep Classifier [100.63114424262234]
Interpretable Deep (IDC) learns the nearest source samples of a target sample as evidence upon which the classifier makes the decision.
Our IDC leads to a more explainable model with almost no accuracy degradation and effectively calibrates classification for optimum reject options.
arXiv Detail & Related papers (2022-11-15T15:58:56Z) - T-NER: An All-Round Python Library for Transformer-based Named Entity
Recognition [9.928025283928282]
T-NER is a Python library for NER LM finetuning.
We show the potential of the library by compiling nine public NER datasets into a unified format.
To facilitate future research, we also release all our LM checkpoints via the Hugging Face model hub.
arXiv Detail & Related papers (2022-09-09T15:00:38Z) - Unsupervised Domain Adaptive Learning via Synthetic Data for Person
Re-identification [101.1886788396803]
Person re-identification (re-ID) has gained more and more attention due to its widespread applications in video surveillance.
Unfortunately, the mainstream deep learning methods still need a large quantity of labeled data to train models.
In this paper, we develop a data collector to automatically generate synthetic re-ID samples in a computer game, and construct a data labeler to simultaneously annotate them.
arXiv Detail & Related papers (2021-09-12T15:51:41Z) - Combining Feature and Instance Attribution to Detect Artifacts [62.63504976810927]
We propose methods to facilitate identification of training data artifacts.
We show that this proposed training-feature attribution approach can be used to uncover artifacts in training data.
We execute a small user study to evaluate whether these methods are useful to NLP researchers in practice.
arXiv Detail & Related papers (2021-07-01T09:26:13Z) - Bootstrapping Relation Extractors using Syntactic Search by Examples [47.11932446745022]
We propose a process for bootstrapping training datasets which can be performed quickly by non-NLP-experts.
We take advantage of search engines over syntactic-graphs which expose a friendly by-example syntax.
We show that the resulting models are competitive with models trained on manually annotated data and on data obtained from distant supervision.
arXiv Detail & Related papers (2021-02-09T18:17:59Z) - Relation-Guided Representation Learning [53.60351496449232]
We propose a new representation learning method that explicitly models and leverages sample relations.
Our framework well preserves the relations between samples.
By seeking to embed samples into subspace, we show that our method can address the large-scale and out-of-sample problem.
arXiv Detail & Related papers (2020-07-11T10:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.