SLK-NER: Exploiting Second-order Lexicon Knowledge for Chinese NER
- URL: http://arxiv.org/abs/2007.08416v1
- Date: Thu, 16 Jul 2020 15:53:02 GMT
- Title: SLK-NER: Exploiting Second-order Lexicon Knowledge for Chinese NER
- Authors: Dou Hu and Lingwei Wei
- Abstract summary: We present new insight into second-order lexicon knowledge (SLK) of each character in the sentence to provide more lexical word information.
The proposed model can exploit more discernible lexical words information with the help of global context.
- Score: 8.122270502556374
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although character-based models using lexicon have achieved promising results
for Chinese named entity recognition (NER) task, some lexical words would
introduce erroneous information due to wrongly matched words. Existing
researches proposed many strategies to integrate lexicon knowledge. However,
they performed with simple first-order lexicon knowledge, which provided
insufficient word information and still faced the challenge of matched word
boundary conflicts; or explored the lexicon knowledge with graph where
higher-order information introducing negative words may disturb the
identification. To alleviate the above limitations, we present new insight into
second-order lexicon knowledge (SLK) of each character in the sentence to
provide more lexical word information including semantic and word boundary
features. Based on these, we propose a SLK-based model with a novel strategy to
integrate the above lexicon knowledge. The proposed model can exploit more
discernible lexical words information with the help of global context.
Experimental results on three public datasets demonstrate the validity of SLK.
The proposed model achieves more excellent performance than the
state-of-the-art comparison methods.
Related papers
- Talking the Talk Does Not Entail Walking the Walk: On the Limits of Large Language Models in Lexical Entailment Recognition [3.8623569699070357]
This work investigates the capabilities of eight Large Language Models in recognizing lexical entailment relations among verbs.
Our findings unveil that the models can tackle the lexical entailment recognition task with moderately good performance.
arXiv Detail & Related papers (2024-06-21T06:30:16Z) - Incorporating Lexical and Syntactic Knowledge for Unsupervised Cross-Lingual Transfer [4.944761231728674]
We present a novel framework called "Lexicon-Syntax Enhanced Multilingual BERT"
We use Multilingual BERT as the base model and employ two techniques to enhance its learning capabilities.
Our experimental results demonstrate this framework can consistently outperform all baselines of zero-shot cross-lingual transfer.
arXiv Detail & Related papers (2024-04-25T14:10:52Z) - Translate to Disambiguate: Zero-shot Multilingual Word Sense
Disambiguation with Pretrained Language Models [67.19567060894563]
Pretrained Language Models (PLMs) learn rich cross-lingual knowledge and can be finetuned to perform well on diverse tasks.
We present a new study investigating how well PLMs capture cross-lingual word sense with Contextual Word-Level Translation (C-WLT)
We find that as the model size increases, PLMs encode more cross-lingual word sense knowledge and better use context to improve WLT performance.
arXiv Detail & Related papers (2023-04-26T19:55:52Z) - The Short Text Matching Model Enhanced with Knowledge via Contrastive
Learning [8.350445155753167]
This paper proposes a short Text Matching model that combines contrastive learning and external knowledge.
To avoid noise, we use keywords as the main semantics of the original sentence to retrieve corresponding knowledge words in the knowledge base.
Our designed model achieves state-of-the-art performance on two publicly available Chinese Text Matching datasets.
arXiv Detail & Related papers (2023-04-08T03:24:05Z) - Learning from the Dictionary: Heterogeneous Knowledge Guided Fine-tuning
for Chinese Spell Checking [32.16787396943434]
Chinese Spell Checking (CSC) aims to detect and correct Chinese spelling errors.
Recent researches start from the pretrained knowledge of language models and take multimodal information into CSC models to improve the performance.
We propose the LEAD framework, which renders the CSC model to learn heterogeneous knowledge from the dictionary in terms of phonetics, vision, and meaning.
arXiv Detail & Related papers (2022-10-19T06:31:34Z) - Always Keep your Target in Mind: Studying Semantics and Improving
Performance of Neural Lexical Substitution [124.99894592871385]
We present a large-scale comparative study of lexical substitution methods employing both old and most recent language models.
We show that already competitive results achieved by SOTA LMs/MLMs can be further substantially improved if information about the target word is injected properly.
arXiv Detail & Related papers (2022-06-07T16:16:19Z) - Exposing Cross-Lingual Lexical Knowledge from Multilingual Sentence
Encoders [85.80950708769923]
We probe multilingual language models for the amount of cross-lingual lexical knowledge stored in their parameters, and compare them against the original multilingual LMs.
We also devise a novel method to expose this knowledge by additionally fine-tuning multilingual models.
We report substantial gains on standard benchmarks.
arXiv Detail & Related papers (2022-04-30T13:23:16Z) - DyLex: Incorporating Dynamic Lexicons into BERT for Sequence Labeling [49.3379730319246]
We propose DyLex, a plug-in lexicon incorporation approach for BERT based sequence labeling tasks.
We adopt word-agnostic tag embeddings to avoid re-training the representation while updating the lexicon.
Finally, we introduce a col-wise attention based knowledge fusion mechanism to guarantee the pluggability of the proposed framework.
arXiv Detail & Related papers (2021-09-18T03:15:49Z) - LexSubCon: Integrating Knowledge from Lexical Resources into Contextual
Embeddings for Lexical Substitution [76.615287796753]
We introduce LexSubCon, an end-to-end lexical substitution framework based on contextual embedding models.
This is achieved by combining contextual information with knowledge from structured lexical resources.
Our experiments show that LexSubCon outperforms previous state-of-the-art methods on LS07 and CoInCo benchmark datasets.
arXiv Detail & Related papers (2021-07-11T21:25:56Z) - Enhanced word embeddings using multi-semantic representation through
lexical chains [1.8199326045904998]
We propose two novel algorithms, called Flexible Lexical Chain II and Fixed Lexical Chain II.
These algorithms combine the semantic relations derived from lexical chains, prior knowledge from lexical databases, and the robustness of the distributional hypothesis in word embeddings as building blocks forming a single system.
Our results show the integration between lexical chains and word embeddings representations sustain state-of-the-art results, even against more complex systems.
arXiv Detail & Related papers (2021-01-22T09:43:33Z) - Probing Pretrained Language Models for Lexical Semantics [76.73599166020307]
We present a systematic empirical analysis across six typologically diverse languages and five different lexical tasks.
Our results indicate patterns and best practices that hold universally, but also point to prominent variations across languages and tasks.
arXiv Detail & Related papers (2020-10-12T14:24:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.