Multi-sense embeddings through a word sense disambiguation process
- URL: http://arxiv.org/abs/2101.08700v1
- Date: Thu, 21 Jan 2021 16:22:34 GMT
- Title: Multi-sense embeddings through a word sense disambiguation process
- Authors: Terry Ruas, William Grosky, Aiko Aizawa
- Abstract summary: Most Suitable Sense.
(MSSA) disambiguates and annotates each word by its specific sense, considering the semantic effects of its context.
We test our approach on six different benchmarks for the word similarity task, showing that our approach can produce state-of-the-art results.
- Score: 2.2344764434954256
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Natural Language Understanding has seen an increasing number of publications
in the last few years, especially after robust word embeddings models became
prominent, when they proved themselves able to capture and represent semantic
relationships from massive amounts of data. Nevertheless, traditional models
often fall short in intrinsic issues of linguistics, such as polysemy and
homonymy. Any expert system that makes use of natural language in its core, can
be affected by a weak semantic representation of text, resulting in inaccurate
outcomes based on poor decisions. To mitigate such issues, we propose a novel
approach called Most Suitable Sense Annotation (MSSA), that disambiguates and
annotates each word by its specific sense, considering the semantic effects of
its context. Our approach brings three main contributions to the semantic
representation scenario: (i) an unsupervised technique that disambiguates and
annotates words by their senses, (ii) a multi-sense embeddings model that can
be extended to any traditional word embeddings algorithm, and (iii) a recurrent
methodology that allows our models to be re-used and their representations
refined. We test our approach on six different benchmarks for the word
similarity task, showing that our approach can produce state-of-the-art results
and outperforms several more complex state-of-the-art systems.
Related papers
- How Well Do Text Embedding Models Understand Syntax? [50.440590035493074]
The ability of text embedding models to generalize across a wide range of syntactic contexts remains under-explored.
Our findings reveal that existing text embedding models have not sufficiently addressed these syntactic understanding challenges.
We propose strategies to augment the generalization ability of text embedding models in diverse syntactic scenarios.
arXiv Detail & Related papers (2023-11-14T08:51:00Z) - A Comprehensive Empirical Evaluation of Existing Word Embedding
Approaches [5.065947993017158]
We present the characteristics of existing word embedding approaches and analyze them with regard to many classification tasks.
Traditional approaches mostly use matrix factorization to produce word representations, and they are not able to capture the semantic and syntactic regularities of the language very well.
On the other hand, Neural-network-based approaches can capture sophisticated regularities of the language and preserve the word relationships in the generated word representations.
arXiv Detail & Related papers (2023-03-13T15:34:19Z) - Contextualized language models for semantic change detection: lessons
learned [4.436724861363513]
We present a qualitative analysis of the outputs of contextualized embedding-based methods for detecting diachronic semantic change.
Our findings show that contextualized methods can often predict high change scores for words which are not undergoing any real diachronic semantic shift.
Our conclusion is that pre-trained contextualized language models are prone to confound changes in lexicographic senses and changes in contextual variance.
arXiv Detail & Related papers (2022-08-31T23:35:24Z) - Idiomatic Expression Identification using Semantic Compatibility [8.355785779504869]
We study the task of detecting whether a sentence has an idiomatic expression and localizing it.
We propose a multi-stage neural architecture with the attention flow mechanism for identifying these expressions.
A salient feature of the model is its ability to identify idioms unseen during training with gains from 1.4% to 30.8% over competitive baselines.
arXiv Detail & Related papers (2021-10-19T15:44:28Z) - Contextualized Semantic Distance between Highly Overlapped Texts [85.1541170468617]
Overlapping frequently occurs in paired texts in natural language processing tasks like text editing and semantic similarity evaluation.
This paper aims to address the issue with a mask-and-predict strategy.
We take the words in the longest common sequence as neighboring words and use masked language modeling (MLM) to predict the distributions on their positions.
Experiments on Semantic Textual Similarity show NDD to be more sensitive to various semantic differences, especially on highly overlapped paired texts.
arXiv Detail & Related papers (2021-10-04T03:59:15Z) - Discrete representations in neural models of spoken language [56.29049879393466]
We compare the merits of four commonly used metrics in the context of weakly supervised models of spoken language.
We find that the different evaluation metrics can give inconsistent results.
arXiv Detail & Related papers (2021-05-12T11:02:02Z) - Language Models for Lexical Inference in Context [4.581468205348204]
Lexical inference in context (LIiC) is the task of recognizing textual entailment between two very similar sentences.
We formulate and evaluate the first approaches based on pretrained language models (LMs) for this task.
All our approaches outperform the previous state of the art, showing the potential of pretrained LMs for LIiC.
arXiv Detail & Related papers (2021-02-10T09:08:22Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Analysing Lexical Semantic Change with Contextualised Word
Representations [7.071298726856781]
We propose a novel method that exploits the BERT neural language model to obtain representations of word usages.
We create a new evaluation dataset and show that the model representations and the detected semantic shifts are positively correlated with human judgements.
arXiv Detail & Related papers (2020-04-29T12:18:14Z) - How Far are We from Effective Context Modeling? An Exploratory Study on
Semantic Parsing in Context [59.13515950353125]
We present a grammar-based decoding semantic parsing and adapt typical context modeling methods on top of it.
We evaluate 13 context modeling methods on two large cross-domain datasets, and our best model achieves state-of-the-art performances.
arXiv Detail & Related papers (2020-02-03T11:28:10Z) - Multiplex Word Embeddings for Selectional Preference Acquisition [70.33531759861111]
We propose a multiplex word embedding model, which can be easily extended according to various relations among words.
Our model can effectively distinguish words with respect to different relations without introducing unnecessary sparseness.
arXiv Detail & Related papers (2020-01-09T04:47:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.