CrossNER: Evaluating Cross-Domain Named Entity Recognition
- URL: http://arxiv.org/abs/2012.04373v2
- Date: Sun, 13 Dec 2020 07:43:16 GMT
- Title: CrossNER: Evaluating Cross-Domain Named Entity Recognition
- Authors: Zihan Liu, Yan Xu, Tiezheng Yu, Wenliang Dai, Ziwei Ji, Samuel
Cahyawijaya, Andrea Madotto, Pascale Fung
- Abstract summary: Cross-domain named entity recognition models are able to cope with the scarcity issue of NER samples in target domains.
Most of the existing NER benchmarks lack domain-specialized entity types or do not focus on a certain domain, leading to a less effective cross-domain evaluation.
We introduce a cross-domain NER dataset (CrossNER), a fully-labeled collection of NER data spanning over five diverse domains with specialized entity categories for different domains.
- Score: 47.9831214875796
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cross-domain named entity recognition (NER) models are able to cope with the
scarcity issue of NER samples in target domains. However, most of the existing
NER benchmarks lack domain-specialized entity types or do not focus on a
certain domain, leading to a less effective cross-domain evaluation. To address
these obstacles, we introduce a cross-domain NER dataset (CrossNER), a
fully-labeled collection of NER data spanning over five diverse domains with
specialized entity categories for different domains. Additionally, we also
provide a domain-related corpus since using it to continue pre-training
language models (domain-adaptive pre-training) is effective for the domain
adaptation. We then conduct comprehensive experiments to explore the
effectiveness of leveraging different levels of the domain corpus and
pre-training strategies to do domain-adaptive pre-training for the cross-domain
task. Results show that focusing on the fractional corpus containing
domain-specialized entities and utilizing a more challenging pre-training
strategy in domain-adaptive pre-training are beneficial for the NER domain
adaptation, and our proposed method can consistently outperform existing
cross-domain NER baselines. Nevertheless, experiments also illustrate the
challenge of this cross-domain NER task. We hope that our dataset and baselines
will catalyze research in the NER domain adaptation area. The code and data are
available at https://github.com/zliucr/CrossNER.
Related papers
- Unsupervised Domain Adaptation for Anatomical Landmark Detection [5.070344284426738]
We propose a novel framework for anatomical landmark detection under the setting of unsupervised domain adaptation (UDA)
The framework leverages self-training and domain adversarial learning to address the domain gap during adaptation.
Our experiments on cephalometric and lung landmark detection show the effectiveness of the method, which reduces the domain gap by a large margin and outperforms other UDA methods consistently.
arXiv Detail & Related papers (2023-08-25T10:22:13Z) - DAOT: Domain-Agnostically Aligned Optimal Transport for Domain-Adaptive
Crowd Counting [35.83485358725357]
Domain adaptation is commonly employed in crowd counting to bridge the domain gaps between different datasets.
Existing domain adaptation methods tend to focus on inter-dataset differences while overlooking the intra-differences within the same dataset.
We propose a Domain-agnostically Aligned Optimal Transport (DAOT) strategy that aligns domain-agnostic factors between domains.
arXiv Detail & Related papers (2023-08-10T02:59:40Z) - One Model for All Domains: Collaborative Domain-Prefix Tuning for
Cross-Domain NER [92.79085995361098]
Cross-domain NER is a challenging task to address the low-resource problem in practical scenarios.
Previous solutions mainly obtain a NER model by pre-trained language models (PLMs) with data from a rich-resource domain and adapt it to the target domain.
We introduce Collaborative Domain-Prefix Tuning for cross-domain NER based on text-to-text generative PLMs.
arXiv Detail & Related papers (2023-01-25T05:16:43Z) - Cross-domain Contrastive Learning for Unsupervised Domain Adaptation [108.63914324182984]
Unsupervised domain adaptation (UDA) aims to transfer knowledge learned from a fully-labeled source domain to a different unlabeled target domain.
We build upon contrastive self-supervised learning to align features so as to reduce the domain discrepancy between training and testing sets.
arXiv Detail & Related papers (2021-06-10T06:32:30Z) - AFAN: Augmented Feature Alignment Network for Cross-Domain Object
Detection [90.18752912204778]
Unsupervised domain adaptation for object detection is a challenging problem with many real-world applications.
We propose a novel augmented feature alignment network (AFAN) which integrates intermediate domain image generation and domain-adversarial training.
Our approach significantly outperforms the state-of-the-art methods on standard benchmarks for both similar and dissimilar domain adaptations.
arXiv Detail & Related papers (2021-06-10T05:01:20Z) - Disentanglement-based Cross-Domain Feature Augmentation for Effective
Unsupervised Domain Adaptive Person Re-identification [87.72851934197936]
Unsupervised domain adaptive (UDA) person re-identification (ReID) aims to transfer the knowledge from the labeled source domain to the unlabeled target domain for person matching.
One challenge is how to generate target domain samples with reliable labels for training.
We propose a Disentanglement-based Cross-Domain Feature Augmentation strategy.
arXiv Detail & Related papers (2021-03-25T15:28:41Z) - Cross-Domain Grouping and Alignment for Domain Adaptive Semantic
Segmentation [74.3349233035632]
Existing techniques to adapt semantic segmentation networks across the source and target domains within deep convolutional neural networks (CNNs) do not consider an inter-class variation within the target domain itself or estimated category.
We introduce a learnable clustering module, and a novel domain adaptation framework called cross-domain grouping and alignment.
Our method consistently boosts the adaptation performance in semantic segmentation, outperforming the state-of-the-arts on various domain adaptation settings.
arXiv Detail & Related papers (2020-12-15T11:36:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.