BOS at LSCDiscovery: Lexical Substitution for Interpretable Lexical
Semantic Change Detection
- URL: http://arxiv.org/abs/2206.11865v1
- Date: Tue, 7 Jun 2022 11:40:29 GMT
- Title: BOS at LSCDiscovery: Lexical Substitution for Interpretable Lexical
Semantic Change Detection
- Authors: Artem Kudisov and Nikolay Arefyev
- Abstract summary: We propose a solution for the LSCDiscovery shared task on Lexical Semantic Change Detection in Spanish.
Our approach is based on generating lexical substitutes that describe old and new senses of a given word.
- Score: 0.48733623015338234
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a solution for the LSCDiscovery shared task on Lexical Semantic
Change Detection in Spanish. Our approach is based on generating lexical
substitutes that describe old and new senses of a given word. This approach
achieves the second best result in sense loss and sense gain detection
subtasks. By observing those substitutes that are specific for only one time
period, one can understand which senses were obtained or lost. This allows
providing more detailed information about semantic change to the user and makes
our method interpretable.
Related papers
- ParaLS: Lexical Substitution via Pretrained Paraphraser [18.929859707202517]
This study explores how to generate the substitute candidates from a paraphraser.
We propose two simple decoding strategies that focus on the variations of the target word during decoding.
arXiv Detail & Related papers (2023-05-14T12:49:16Z) - Are Synonym Substitution Attacks Really Synonym Substitution Attacks? [80.81532239566992]
We show that four widely used word substitution methods generate a large fraction of invalid substitution words that are ungrammatical or do not preserve the original sentence's semantics.
Next, we show that the semantic and grammatical constraints used in SSAs for detecting invalid word replacements are highly insufficient in detecting invalid adversarial samples.
arXiv Detail & Related papers (2022-10-06T12:01:50Z) - Always Keep your Target in Mind: Studying Semantics and Improving
Performance of Neural Lexical Substitution [124.99894592871385]
We present a large-scale comparative study of lexical substitution methods employing both old and most recent language models.
We show that already competitive results achieved by SOTA LMs/MLMs can be further substantially improved if information about the target word is injected properly.
arXiv Detail & Related papers (2022-06-07T16:16:19Z) - Large Scale Substitution-based Word Sense Induction [48.49573297876054]
We present a word-sense induction method based on pre-trained masked language models (MLMs), which can cheaply scale to large vocabularies and large corpora.
The result is a corpus which is sense-tagged according to a corpus-derived sense inventory and where each sense is associated with indicative words.
Evaluation on English Wikipedia that was sense-tagged using our method shows that both the induced senses, and the per-instance sense assignment, are of high quality even compared to WSD methods, such as Babelfy.
arXiv Detail & Related papers (2021-10-14T19:40:37Z) - LexSubCon: Integrating Knowledge from Lexical Resources into Contextual
Embeddings for Lexical Substitution [76.615287796753]
We introduce LexSubCon, an end-to-end lexical substitution framework based on contextual embedding models.
This is achieved by combining contextual information with knowledge from structured lexical resources.
Our experiments show that LexSubCon outperforms previous state-of-the-art methods on LS07 and CoInCo benchmark datasets.
arXiv Detail & Related papers (2021-07-11T21:25:56Z) - Fake it Till You Make it: Self-Supervised Semantic Shifts for
Monolingual Word Embedding Tasks [58.87961226278285]
We propose a self-supervised approach to model lexical semantic change.
We show that our method can be used for the detection of semantic change with any alignment method.
We illustrate the utility of our techniques using experimental results on three different datasets.
arXiv Detail & Related papers (2021-01-30T18:59:43Z) - UWB at SemEval-2020 Task 1: Lexical Semantic Change Detection [1.2599533416395767]
We examine semantic differences between specific words in two corpora, chosen from different time periods, for English, German, Latin, and Swedish.
Our method was created for the SemEval 2020 Task 1: textitUnsupervised Lexical Semantic Change Detection.
arXiv Detail & Related papers (2020-11-30T10:47:45Z) - UoB at SemEval-2020 Task 1: Automatic Identification of Novel Word
Senses [0.6980076213134383]
This paper presents an approach to lexical semantic change detection based on Bayesian word sense induction suitable for novel word sense identification.
The same approach is also applied to a corpus gleaned from 15 years of Twitter data, the results of which are then used to identify words which may be instances of slang.
arXiv Detail & Related papers (2020-10-18T19:27:06Z) - SemEval-2020 Task 1: Unsupervised Lexical Semantic Change Detection [10.606357227329822]
Evaluation is currently the most pressing problem in Lexical Semantic Change detection.
No gold standards are available to the community, which hinders progress.
We present the results of the first shared task that addresses this gap.
arXiv Detail & Related papers (2020-07-22T14:37:42Z) - Unsupervised Embedding-based Detection of Lexical Semantic Changes [1.7403133838762452]
This paper describes EmbLexChange, a system introduced by the "Life-Language" team for SemEval-2020 Task 1.
EmmLexChange is defined as the divergence between the embedding based profiles of word w in the source and the target domains.
We show that using a resampling framework for the selection of reference words, we can reliably detect lexical-semantic changes in English, German, Swedish, and Latin.
arXiv Detail & Related papers (2020-05-16T13:05:47Z) - Moving Down the Long Tail of Word Sense Disambiguation with
Gloss-Informed Biencoders [79.38278330678965]
A major obstacle in Word Sense Disambiguation (WSD) is that word senses are not uniformly distributed.
We propose a bi-encoder model that independently embeds (1) the target word with its surrounding context and (2) the dictionary definition, or gloss, of each sense.
arXiv Detail & Related papers (2020-05-06T04:21:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.