FactMix: Using a Few Labeled In-domain Examples to Generalize to
Cross-domain Named Entity Recognition
- URL: http://arxiv.org/abs/2208.11464v3
- Date: Tue, 26 Dec 2023 03:48:59 GMT
- Title: FactMix: Using a Few Labeled In-domain Examples to Generalize to
Cross-domain Named Entity Recognition
- Authors: Linyi Yang, Lifan Yuan, Leyang Cui, Wenyang Gao, Yue Zhang
- Abstract summary: This paper proposes a two-step rationale-centric data augmentation method to improve the model's generalization ability.
Results on several datasets show that our model-agnostic method significantly improves the performance of cross-domain NER tasks.
- Score: 42.32824906747491
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Few-shot Named Entity Recognition (NER) is imperative for entity tagging in
limited resource domains and thus received proper attention in recent years.
Existing approaches for few-shot NER are evaluated mainly under in-domain
settings. In contrast, little is known about how these inherently faithful
models perform in cross-domain NER using a few labeled in-domain examples. This
paper proposes a two-step rationale-centric data augmentation method to improve
the model's generalization ability. Results on several datasets show that our
model-agnostic method significantly improves the performance of cross-domain
NER tasks compared to previous state-of-the-art methods, including the data
augmentation and prompt-tuning methods. Our codes are available at
https://github.com/lifan-yuan/FactMix.
Related papers
- Improving Few-Shot Cross-Domain Named Entity Recognition by Instruction Tuning a Word-Embedding based Retrieval Augmented Large Language Model [0.0]
Few-Shot Cross-Domain NER is a process of leveraging knowledge from data-rich source domains to perform entity recognition on data scarce target domains.
We propose IF-WRANER, a retrieval augmented large language model for Named Entity Recognition.
arXiv Detail & Related papers (2024-11-01T08:57:29Z) - Multi-Domain Long-Tailed Learning by Augmenting Disentangled
Representations [80.76164484820818]
There is an inescapable long-tailed class-imbalance issue in many real-world classification problems.
We study this multi-domain long-tailed learning problem and aim to produce a model that generalizes well across all classes and domains.
Built upon a proposed selective balanced sampling strategy, TALLY achieves this by mixing the semantic representation of one example with the domain-associated nuisances of another.
arXiv Detail & Related papers (2022-10-25T21:54:26Z) - Low-confidence Samples Matter for Domain Adaptation [47.552605279925736]
Domain adaptation (DA) aims to transfer knowledge from a label-rich source domain to a related but label-scarce target domain.
We propose a novel contrastive learning method by processing low-confidence samples.
We evaluate the proposed method in both unsupervised and semi-supervised DA settings.
arXiv Detail & Related papers (2022-02-06T15:45:45Z) - CONTaiNER: Few-Shot Named Entity Recognition via Contrastive Learning [11.289324473201614]
Existing approaches only learn class-specific semantic features and intermediate representations from source domains.
We present CONTaiNER, a novel contrastive learning technique that optimize the inter-token distribution distance for Few-Shot NER.
Our experiments in several traditional test domains (OntoNotes, CoNLL'03, WNUT '17, GUM) and a new large scale Few-Shot NER dataset (Few-NERD) demonstrate that CONTaiNER outperforms previous methods by 3%-13% absolute F1 points.
arXiv Detail & Related papers (2021-09-15T21:41:16Z) - Enhancing the Generalization for Intent Classification and Out-of-Domain
Detection in SLU [70.44344060176952]
Intent classification is a major task in spoken language understanding (SLU)
Recent works have shown that using extra data and labels can improve the OOD detection performance.
This paper proposes to train a model with only IND data while supporting both IND intent classification and OOD detection.
arXiv Detail & Related papers (2021-06-28T08:27:38Z) - Batch Normalization Embeddings for Deep Domain Generalization [50.51405390150066]
Domain generalization aims at training machine learning models to perform robustly across different and unseen domains.
We show a significant increase in classification accuracy over current state-of-the-art techniques on popular domain generalization benchmarks.
arXiv Detail & Related papers (2020-11-25T12:02:57Z) - Named Entity Recognition without Labelled Data: A Weak Supervision
Approach [23.05371427663683]
This paper presents a simple but powerful approach to learn NER models in the absence of labelled data through weak supervision.
The approach relies on a broad spectrum of labelling functions to automatically annotate texts from the target domain.
A sequence labelling model can finally be trained on the basis of this unified annotation.
arXiv Detail & Related papers (2020-04-30T12:29:55Z) - Towards Fair Cross-Domain Adaptation via Generative Learning [50.76694500782927]
Domain Adaptation (DA) targets at adapting a model trained over the well-labeled source domain to the unlabeled target domain lying in different distributions.
We develop a novel Generative Few-shot Cross-domain Adaptation (GFCA) algorithm for fair cross-domain classification.
arXiv Detail & Related papers (2020-03-04T23:25:09Z) - Zero-Resource Cross-Domain Named Entity Recognition [68.83177074227598]
Existing models for cross-domain named entity recognition rely on numerous unlabeled corpus or labeled NER training data in target domains.
We propose a cross-domain NER model that does not use any external resources.
arXiv Detail & Related papers (2020-02-14T09:04:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.