IRB-NLP at SemEval-2022 Task 1: Exploring the Relationship Between Words
and Their Semantic Representations
- URL: http://arxiv.org/abs/2205.06840v1
- Date: Fri, 13 May 2022 18:15:20 GMT
- Title: IRB-NLP at SemEval-2022 Task 1: Exploring the Relationship Between Words
and Their Semantic Representations
- Authors: Damir Koren\v{c}i\'c, Ivan Grubi\v{s}i\'c
- Abstract summary: We present our findings based on the descriptive, exploratory, and predictive data analysis conducted on the CODWOE dataset.
We give a detailed overview of the systems that we designed for Definition Modeling and Reverse Dictionary tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: What is the relation between a word and its description, or a word and its
embedding? Both descriptions and embeddings are semantic representations of
words. But, what information from the original word remains in these
representations? Or more importantly, which information about a word do these
two representations share? Definition Modeling and Reverse Dictionary are two
opposite learning tasks that address these questions. The goal of the
Definition Modeling task is to investigate the power of information laying
inside a word embedding to express the meaning of the word in a humanly
understandable way -- as a dictionary definition. Conversely, the Reverse
Dictionary task explores the ability to predict word embeddings directly from
its definition. In this paper, by tackling these two tasks, we are exploring
the relationship between words and their semantic representations. We present
our findings based on the descriptive, exploratory, and predictive data
analysis conducted on the CODWOE dataset. We give a detailed overview of the
systems that we designed for Definition Modeling and Reverse Dictionary tasks,
and that achieved top scores on SemEval-2022 CODWOE challenge in several
subtasks. We hope that our experimental results concerning the predictive
models and the data analyses we provide will prove useful in future
explorations of word representations and their relationships.
Related papers
- Identifying and interpreting non-aligned human conceptual
representations using language modeling [0.0]
We show that congenital blindness induces conceptual reorganization in both a-modal and sensory-related verbal domains.
We find that blind individuals more strongly associate social and cognitive meanings to verbs related to motion.
For some verbs, representations of blind and sighted are highly similar.
arXiv Detail & Related papers (2024-03-10T13:02:27Z) - Domain Embeddings for Generating Complex Descriptions of Concepts in
Italian Language [65.268245109828]
We propose a Distributional Semantic resource enriched with linguistic and lexical information extracted from electronic dictionaries.
The resource comprises 21 domain-specific matrices, one comprehensive matrix, and a Graphical User Interface.
Our model facilitates the generation of reasoned semantic descriptions of concepts by selecting matrices directly associated with concrete conceptual knowledge.
arXiv Detail & Related papers (2024-02-26T15:04:35Z) - Quantifying the redundancy between prosody and text [67.07817268372743]
We use large language models to estimate how much information is redundant between prosody and the words themselves.
We find a high degree of redundancy between the information carried by the words and prosodic information across several prosodic features.
Still, we observe that prosodic features can not be fully predicted from text, suggesting that prosody carries information above and beyond the words.
arXiv Detail & Related papers (2023-11-28T21:15:24Z) - "Definition Modeling: To model definitions." Generating Definitions With
Little to No Semantics [0.4061135251278187]
We present evidence that the task may not involve as much semantics as one might expect.
We show how an earlier model from the literature is both rather insensitive to semantic aspects such as explicit polysemy.
arXiv Detail & Related papers (2023-06-14T11:08:38Z) - A Unified Model for Reverse Dictionary and Definition Modelling [7.353994554197792]
We train a dual-way neural dictionary to guess words from definitions (reverse dictionary) and produce definitions given words (definition modelling)
Our method learns the two tasks simultaneously, and handles unknown words via embeddings.
It casts a word or a definition to the same representation space through a shared layer, then generates the other form from there, in a multi-task fashion.
arXiv Detail & Related papers (2022-05-09T23:52:39Z) - Connect-the-Dots: Bridging Semantics between Words and Definitions via
Aligning Word Sense Inventories [47.03271152494389]
Word Sense Disambiguation aims to automatically identify the exact meaning of one word according to its context.
Existing supervised models struggle to make correct predictions on rare word senses due to limited training data.
We propose a gloss alignment algorithm that can align definition sentences with the same meaning from different sense inventories to collect rich lexical knowledge.
arXiv Detail & Related papers (2021-10-27T00:04:33Z) - Understanding Synonymous Referring Expressions via Contrastive Features [105.36814858748285]
We develop an end-to-end trainable framework to learn contrastive features on the image and object instance levels.
We conduct extensive experiments to evaluate the proposed algorithm on several benchmark datasets.
arXiv Detail & Related papers (2021-04-20T17:56:24Z) - Words aren't enough, their order matters: On the Robustness of Grounding
Visual Referring Expressions [87.33156149634392]
We critically examine RefCOg, a standard benchmark for visual referring expression recognition.
We show that 83.7% of test instances do not require reasoning on linguistic structure.
We propose two methods, one based on contrastive learning and the other based on multi-task learning, to increase the robustness of ViLBERT.
arXiv Detail & Related papers (2020-05-04T17:09:15Z) - CIRCE at SemEval-2020 Task 1: Ensembling Context-Free and
Context-Dependent Word Representations [0.0]
We present an ensemble model that makes predictions based on context-free and context-dependent word representations.
The key findings are that (1) context-free word representations are a powerful and robust baseline, (2) a sentence classification objective can be used to obtain useful context-dependent word representations, and (3) combining those representations increases performance on some datasets while decreasing performance on others.
arXiv Detail & Related papers (2020-04-30T13:18:29Z) - Lexical Sememe Prediction using Dictionary Definitions by Capturing
Local Semantic Correspondence [94.79912471702782]
Sememes, defined as the minimum semantic units of human languages, have been proven useful in many NLP tasks.
We propose a Sememe Correspondence Pooling (SCorP) model, which is able to capture this kind of matching to predict sememes.
We evaluate our model and baseline methods on a famous sememe KB HowNet and find that our model achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-01-16T17:30:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.