Wiki to Automotive: Understanding the Distribution Shift and its impact
on Named Entity Recognition
- URL: http://arxiv.org/abs/2112.00283v1
- Date: Wed, 1 Dec 2021 05:13:47 GMT
- Title: Wiki to Automotive: Understanding the Distribution Shift and its impact
on Named Entity Recognition
- Authors: Anmol Nayak, Hari Prasad Timmapathini
- Abstract summary: Transfer learning is often unable to replicate the performance of pre-trained models on text of niche domains like Automotive.
We focus on performing the Named Entity Recognition (NER) task as it requires strong lexical, syntactic and semantic understanding by the model.
Fine-tuning the language models with automotive domain text did not make significant improvements to the NER performance.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While transfer learning has become a ubiquitous technique used across Natural
Language Processing (NLP) tasks, it is often unable to replicate the
performance of pre-trained models on text of niche domains like Automotive. In
this paper we aim to understand the main characteristics of the distribution
shift with automotive domain text (describing technical functionalities such as
Cruise Control) and attempt to explain the potential reasons for the gap in
performance. We focus on performing the Named Entity Recognition (NER) task as
it requires strong lexical, syntactic and semantic understanding by the model.
Our experiments with 2 different encoders, namely BERT-Base-Uncased and
SciBERT-Base-Scivocab-Uncased have lead to interesting findings that showed: 1)
The performance of SciBERT is better than BERT when used for automotive domain,
2) Fine-tuning the language models with automotive domain text did not make
significant improvements to the NER performance, 3) The distribution shift is
challenging as it is characterized by lack of repeating contexts, sparseness of
entities, large number of Out-Of-Vocabulary (OOV) words and class overlap due
to domain specific nuances.
Related papers
- Dude: Dual Distribution-Aware Context Prompt Learning For Large Vision-Language Model [27.56988000960972]
We introduce a new framework based on a dual context of both domain-shared and class-specific contexts.
Such dual prompt methods enhance the model's feature representation by joining implicit and explicit factors encoded in Large Language Models.
We also formulate the Unbalanced Optimal Transport (UOT) theory to quantify the relationships between constructed prompts and visual tokens.
arXiv Detail & Related papers (2024-07-05T13:15:29Z) - Incorporating Class-based Language Model for Named Entity Recognition in Factorized Neural Transducer [50.572974726351504]
We propose C-FNT, a novel E2E model that incorporates class-based LMs into FNT.
In C-FNT, the LM score of named entities can be associated with the name class instead of its surface form.
The experimental results show that our proposed C-FNT significantly reduces error in named entities without hurting performance in general word recognition.
arXiv Detail & Related papers (2023-09-14T12:14:49Z) - Domain-oriented Language Pre-training with Adaptive Hybrid Masking and
Optimal Transport Alignment [43.874781718934486]
We provide a general domain-oriented approach to adapt pre-trained language models for different application domains.
To preserve phrase knowledge effectively, we build a domain phrase pool as auxiliary training tool.
We introduce Cross Entity Alignment to leverage entity association as weak supervision to augment the semantic learning of pre-trained models.
arXiv Detail & Related papers (2021-12-01T15:47:01Z) - AutoTriggER: Label-Efficient and Robust Named Entity Recognition with
Auxiliary Trigger Extraction [54.20039200180071]
We present a novel framework to improve NER performance by automatically generating and leveraging entity triggers''
Our framework leverages post-hoc explanation to generate rationales and strengthens a model's prior knowledge using an embedding technique.
AutoTriggER shows strong label-efficiency, is capable of generalizing to unseen entities, and outperforms the RoBERTa-CRF baseline by nearly 0.5 F1 points on average.
arXiv Detail & Related papers (2021-09-10T08:11:56Z) - Domain specific BERT representation for Named Entity Recognition of lab
protocol [0.0]
BERT family seems to work exceptionally well on the downstream task from NER tagging to the range of other linguistic tasks.
In this paper, we are going to illustrate the System for Named Entity Tagging based on Bio-Bert.
arXiv Detail & Related papers (2020-12-21T06:54:38Z) - Neural Entity Linking on Technical Service Tickets [1.3621712165154805]
We show that a neural approach outperforms and complements hand-coded entities, with improvements of about 20% top-1 accuracy.
We also show that a simple sentence-wise encoding (Bi-Encoder) offers a fast yet efficient search in practice.
arXiv Detail & Related papers (2020-05-15T15:47:02Z) - Explicit Domain Adaptation with Loosely Coupled Samples [85.9511585604837]
We propose a transfer learning framework, core of which is learning an explicit mapping between domains.
Due to its interpretability, this is beneficial for safety-critical applications, like autonomous driving.
arXiv Detail & Related papers (2020-04-24T21:23:45Z) - Probing Linguistic Features of Sentence-Level Representations in Neural
Relation Extraction [80.38130122127882]
We introduce 14 probing tasks targeting linguistic properties relevant to neural relation extraction (RE)
We use them to study representations learned by more than 40 different encoder architecture and linguistic feature combinations trained on two datasets.
We find that the bias induced by the architecture and the inclusion of linguistic features are clearly expressed in the probing task performance.
arXiv Detail & Related papers (2020-04-17T09:17:40Z) - Interpretability Analysis for Named Entity Recognition to Understand
System Predictions and How They Can Improve [49.878051587667244]
We examine the performance of several variants of LSTM-CRF architectures for named entity recognition.
We find that context representations do contribute to system performance, but that the main factor driving high performance is learning the name tokens themselves.
We enlist human annotators to evaluate the feasibility of inferring entity types from the context alone and find that, while people are not able to infer the entity type either for the majority of the errors made by the context-only system, there is some room for improvement.
arXiv Detail & Related papers (2020-04-09T14:37:12Z) - Domain Adaption for Knowledge Tracing [65.86619804954283]
We propose a novel adaptable framework, namely knowledge tracing (AKT) to address the DAKT problem.
For the first aspect, we incorporate the educational characteristics (e.g., slip, guess, question texts) based on the deep knowledge tracing (DKT) to obtain a good performed knowledge tracing model.
For the second aspect, we propose and adopt three domain adaptation processes. First, we pre-train an auto-encoder to select useful source instances for target model training.
arXiv Detail & Related papers (2020-01-14T15:04:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.