Lexical Sememe Prediction using Dictionary Definitions by Capturing
Local Semantic Correspondence
- URL: http://arxiv.org/abs/2001.05954v1
- Date: Thu, 16 Jan 2020 17:30:36 GMT
- Title: Lexical Sememe Prediction using Dictionary Definitions by Capturing
Local Semantic Correspondence
- Authors: Jiaju Du, Fanchao Qi, Maosong Sun, Zhiyuan Liu
- Abstract summary: Sememes, defined as the minimum semantic units of human languages, have been proven useful in many NLP tasks.
We propose a Sememe Correspondence Pooling (SCorP) model, which is able to capture this kind of matching to predict sememes.
We evaluate our model and baseline methods on a famous sememe KB HowNet and find that our model achieves state-of-the-art performance.
- Score: 94.79912471702782
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sememes, defined as the minimum semantic units of human languages in
linguistics, have been proven useful in many NLP tasks. Since manual
construction and update of sememe knowledge bases (KBs) are costly, the task of
automatic sememe prediction has been proposed to assist sememe annotation. In
this paper, we explore the approach of applying dictionary definitions to
predicting sememes for unannotated words. We find that sememes of each word are
usually semantically matched to different words in its dictionary definition,
and we name this matching relationship local semantic correspondence.
Accordingly, we propose a Sememe Correspondence Pooling (SCorP) model, which is
able to capture this kind of matching to predict sememes. We evaluate our model
and baseline methods on a famous sememe KB HowNet and find that our model
achieves state-of-the-art performance. Moreover, further quantitative analysis
shows that our model can properly learn the local semantic correspondence
between sememes and words in dictionary definitions, which explains the
effectiveness of our model. The source codes of this paper can be obtained from
https://github.com/thunlp/scorp.
Related papers
- Contextual Dictionary Lookup for Knowledge Graph Completion [32.493168863565465]
Knowledge graph completion (KGC) aims to solve the incompleteness of knowledge graphs (KGs) by predicting missing links from known triples.
Most existing embedding models map each relation into a unique vector, overlooking the specific fine-grained semantics of them under different entities.
We present a novel method utilizing contextual dictionary lookup, enabling conventional embedding models to learn fine-grained semantics of relations in an end-to-end manner.
arXiv Detail & Related papers (2023-06-13T12:13:41Z) - Unsupervised Semantic Variation Prediction using the Distribution of
Sibling Embeddings [17.803726860514193]
Detection of semantic variation of words is an important task for various NLP applications.
We argue that mean representations alone cannot accurately capture such semantic variations.
We propose a method that uses the entire cohort of the contextualised embeddings of the target word.
arXiv Detail & Related papers (2023-05-15T13:58:21Z) - Chinese Word Sense Embedding with SememeWSD and Synonym Set [17.37973450772783]
We propose SememeWSD Synonym (SWSDS) model to assign a different vector to every sense of polysemous words.
We obtain top 10 synonyms of the word sense from OpenHowNet and calculate the average vector of synonyms as the vector of the word sense.
In experiments, We evaluate the SWSDS model on semantic similarity calculation with Gensim's wmdistance method.
arXiv Detail & Related papers (2022-06-29T03:42:03Z) - A Unified Model for Reverse Dictionary and Definition Modelling [7.353994554197792]
We train a dual-way neural dictionary to guess words from definitions (reverse dictionary) and produce definitions given words (definition modelling)
Our method learns the two tasks simultaneously, and handles unknown words via embeddings.
It casts a word or a definition to the same representation space through a shared layer, then generates the other form from there, in a multi-task fashion.
arXiv Detail & Related papers (2022-05-09T23:52:39Z) - Few-Shot Semantic Parsing with Language Models Trained On Code [52.23355024995237]
We find that Codex performs better at semantic parsing than equivalent GPT-3 models.
We find that unlike GPT-3, Codex performs similarly when targeting meaning representations directly, perhaps as meaning representations used in semantic parsing are structured similar to code.
arXiv Detail & Related papers (2021-12-16T08:34:06Z) - Dict-BERT: Enhancing Language Model Pre-training with Dictionary [42.0998323292348]
Pre-trained language models (PLMs) aim to learn universal language representations by conducting self-supervised training tasks on large-scale corpora.
In this work, we focus on enhancing language model pre-training by leveraging definitions of rare words in dictionaries.
We propose two novel self-supervised pre-training tasks on word and sentence-level alignment between input text sequence and rare word definitions.
arXiv Detail & Related papers (2021-10-13T04:29:14Z) - Fake it Till You Make it: Self-Supervised Semantic Shifts for
Monolingual Word Embedding Tasks [58.87961226278285]
We propose a self-supervised approach to model lexical semantic change.
We show that our method can be used for the detection of semantic change with any alignment method.
We illustrate the utility of our techniques using experimental results on three different datasets.
arXiv Detail & Related papers (2021-01-30T18:59:43Z) - SChME at SemEval-2020 Task 1: A Model Ensemble for Detecting Lexical
Semantic Change [58.87961226278285]
This paper describes SChME, a method used in SemEval-2020 Task 1 on unsupervised detection of lexical semantic change.
SChME usesa model ensemble combining signals of distributional models (word embeddings) and wordfrequency models where each model casts a vote indicating the probability that a word sufferedsemantic change according to that feature.
arXiv Detail & Related papers (2020-12-02T23:56:34Z) - NLP-CIC @ DIACR-Ita: POS and Neighbor Based Distributional Models for
Lexical Semantic Change in Diachronic Italian Corpora [62.997667081978825]
We present our systems and findings on unsupervised lexical semantic change for the Italian language.
The task is to determine whether a target word has evolved its meaning with time, only relying on raw-text from two time-specific datasets.
We propose two models representing the target words across the periods to predict the changing words using threshold and voting schemes.
arXiv Detail & Related papers (2020-11-07T11:27:18Z) - Grounded Compositional Outputs for Adaptive Language Modeling [59.02706635250856]
A language model's vocabulary$-$typically selected before training and permanently fixed later$-$affects its size.
We propose a fully compositional output embedding layer for language models.
To our knowledge, the result is the first word-level language model with a size that does not depend on the training vocabulary.
arXiv Detail & Related papers (2020-09-24T07:21:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.