PREDICT: Persian Reverse Dictionary
- URL: http://arxiv.org/abs/2105.00309v1
- Date: Sat, 1 May 2021 17:37:01 GMT
- Title: PREDICT: Persian Reverse Dictionary
- Authors: Arman Malekzadeh and Amin Gheibi and Ali Mohades
- Abstract summary: We compare four different architectures for implementing a Persian reverse dictionary (PREDICT)
We evaluate our models using (phrase,word)Words extracted from the only Persian dictionaries available online.
Experiments show that a model consisting of Long Short-Term Memory (LSTM) units enhanced by an additive attention mechanism is enough to produce suggestions comparable to (or in some cases better than) the word in the original dictionary.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Finding the appropriate words to convey concepts (i.e., lexical access) is
essential for effective communication. Reverse dictionaries fulfill this need
by helping individuals to find the word(s) which could relate to a specific
concept or idea. To the best of our knowledge, this resource has not been
available for the Persian language. In this paper, we compare four different
architectures for implementing a Persian reverse dictionary (PREDICT).
We evaluate our models using (phrase,word) tuples extracted from the only
Persian dictionaries available online, namely Amid, Moein, and Dehkhoda where
the phrase describes the word. Given the phrase, a model suggests the most
relevant word(s) in terms of the ability to convey the concept. The model is
considered to perform well if the correct word is one of its top suggestions.
Our experiments show that a model consisting of Long Short-Term Memory (LSTM)
units enhanced by an additive attention mechanism is enough to produce
suggestions comparable to (or in some cases better than) the word in the
original dictionary. The study also reveals that the model sometimes produces
the synonyms of the word as its output which led us to introduce a new metric
for the evaluation of reverse dictionaries called Synonym Accuracy accounting
for the percentage of times the event of producing the word or a synonym of it
occurs. The assessment of the best model using this new metric also indicates
that at least 62% of the times, it produces an accurate result within the top
100 suggestions.
Related papers
- Biomedical Named Entity Recognition via Dictionary-based Synonym
Generalization [51.89486520806639]
We propose a novel Synonym Generalization (SynGen) framework that recognizes the biomedical concepts contained in the input text using span-based predictions.
We extensively evaluate our approach on a wide range of benchmarks and the results verify that SynGen outperforms previous dictionary-based models by notable margins.
arXiv Detail & Related papers (2023-05-22T14:36:32Z) - Chinese Word Sense Embedding with SememeWSD and Synonym Set [17.37973450772783]
We propose SememeWSD Synonym (SWSDS) model to assign a different vector to every sense of polysemous words.
We obtain top 10 synonyms of the word sense from OpenHowNet and calculate the average vector of synonyms as the vector of the word sense.
In experiments, We evaluate the SWSDS model on semantic similarity calculation with Gensim's wmdistance method.
arXiv Detail & Related papers (2022-06-29T03:42:03Z) - A Unified Model for Reverse Dictionary and Definition Modelling [7.353994554197792]
We train a dual-way neural dictionary to guess words from definitions (reverse dictionary) and produce definitions given words (definition modelling)
Our method learns the two tasks simultaneously, and handles unknown words via embeddings.
It casts a word or a definition to the same representation space through a shared layer, then generates the other form from there, in a multi-task fashion.
arXiv Detail & Related papers (2022-05-09T23:52:39Z) - Dict-BERT: Enhancing Language Model Pre-training with Dictionary [42.0998323292348]
Pre-trained language models (PLMs) aim to learn universal language representations by conducting self-supervised training tasks on large-scale corpora.
In this work, we focus on enhancing language model pre-training by leveraging definitions of rare words in dictionaries.
We propose two novel self-supervised pre-training tasks on word and sentence-level alignment between input text sequence and rare word definitions.
arXiv Detail & Related papers (2021-10-13T04:29:14Z) - Meta-Learning with Variational Semantic Memory for Word Sense
Disambiguation [56.830395467247016]
We propose a model of semantic memory for WSD in a meta-learning setting.
Our model is based on hierarchical variational inference and incorporates an adaptive memory update rule via a hypernetwork.
We show our model advances the state of the art in few-shot WSD, supports effective learning in extremely data scarce scenarios.
arXiv Detail & Related papers (2021-06-05T20:40:01Z) - NLP-CIC @ DIACR-Ita: POS and Neighbor Based Distributional Models for
Lexical Semantic Change in Diachronic Italian Corpora [62.997667081978825]
We present our systems and findings on unsupervised lexical semantic change for the Italian language.
The task is to determine whether a target word has evolved its meaning with time, only relying on raw-text from two time-specific datasets.
We propose two models representing the target words across the periods to predict the changing words using threshold and voting schemes.
arXiv Detail & Related papers (2020-11-07T11:27:18Z) - Grounded Compositional Outputs for Adaptive Language Modeling [59.02706635250856]
A language model's vocabulary$-$typically selected before training and permanently fixed later$-$affects its size.
We propose a fully compositional output embedding layer for language models.
To our knowledge, the result is the first word-level language model with a size that does not depend on the training vocabulary.
arXiv Detail & Related papers (2020-09-24T07:21:14Z) - Comparative Analysis of Word Embeddings for Capturing Word Similarities [0.0]
Distributed language representation has become the most widely used technique for language representation in various natural language processing tasks.
Most of the natural language processing models that are based on deep learning techniques use already pre-trained distributed word representations, commonly called word embeddings.
selecting the appropriate word embeddings is a perplexing task since the projected embedding space is not intuitive to humans.
arXiv Detail & Related papers (2020-05-08T01:16:03Z) - Word Sense Disambiguation for 158 Languages using Word Embeddings Only [80.79437083582643]
Disambiguation of word senses in context is easy for humans, but a major challenge for automatic approaches.
We present a method that takes as input a standard pre-trained word embedding model and induces a fully-fledged word sense inventory.
We use this method to induce a collection of sense inventories for 158 languages on the basis of the original pre-trained fastText word embeddings.
arXiv Detail & Related papers (2020-03-14T14:50:04Z) - Lexical Sememe Prediction using Dictionary Definitions by Capturing
Local Semantic Correspondence [94.79912471702782]
Sememes, defined as the minimum semantic units of human languages, have been proven useful in many NLP tasks.
We propose a Sememe Correspondence Pooling (SCorP) model, which is able to capture this kind of matching to predict sememes.
We evaluate our model and baseline methods on a famous sememe KB HowNet and find that our model achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-01-16T17:30:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.