Towards a Theoretical Understanding of Word and Relation Representation
- URL: http://arxiv.org/abs/2202.00486v1
- Date: Tue, 1 Feb 2022 15:34:58 GMT
- Title: Towards a Theoretical Understanding of Word and Relation Representation
- Authors: Carl Allen
- Abstract summary: Representing words by vectors, or embeddings, enables computational reasoning.
We focus on word embeddings learned from text corpora and knowledge graphs.
- Score: 8.020742121274418
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Representing words by vectors, or embeddings, enables computational reasoning
and is foundational to automating natural language tasks. For example, if word
embeddings of similar words contain similar values, word similarity can be
readily assessed, whereas judging that from their spelling is often impossible
(e.g. cat /feline) and to predetermine and store similarities between all words
is prohibitively time-consuming, memory intensive and subjective. We focus on
word embeddings learned from text corpora and knowledge graphs. Several
well-known algorithms learn word embeddings from text on an unsupervised basis
by learning to predict those words that occur around each word, e.g. word2vec
and GloVe. Parameters of such word embeddings are known to reflect word
co-occurrence statistics, but how they capture semantic meaning has been
unclear. Knowledge graph representation models learn representations both of
entities (words, people, places, etc.) and relations between them, typically by
training a model to predict known facts in a supervised manner. Despite steady
improvements in fact prediction accuracy, little is understood of the latent
structure that enables this.
The limited understanding of how latent semantic structure is encoded in the
geometry of word embeddings and knowledge graph representations makes a
principled means of improving their performance, reliability or
interpretability unclear. To address this:
1. we theoretically justify the empirical observation that particular
geometric relationships between word embeddings learned by algorithms such as
word2vec and GloVe correspond to semantic relations between words; and
2. we extend this correspondence between semantics and geometry to the
entities and relations of knowledge graphs, providing a model for the latent
structure of knowledge graph representation linked to that of word embeddings.
Related papers
- Neighboring Words Affect Human Interpretation of Saliency Explanations [65.29015910991261]
Word-level saliency explanations are often used to communicate feature-attribution in text-based models.
Recent studies found that superficial factors such as word length can distort human interpretation of the communicated saliency scores.
We investigate how the marking of a word's neighboring words affect the explainee's perception of the word's importance in the context of a saliency explanation.
arXiv Detail & Related papers (2023-05-04T09:50:25Z) - How Do Transformers Learn Topic Structure: Towards a Mechanistic
Understanding [56.222097640468306]
We provide mechanistic understanding of how transformers learn "semantic structure"
We show, through a combination of mathematical analysis and experiments on Wikipedia data, that the embedding layer and the self-attention layer encode the topical structure.
arXiv Detail & Related papers (2023-03-07T21:42:17Z) - Variational Cross-Graph Reasoning and Adaptive Structured Semantics
Learning for Compositional Temporal Grounding [143.5927158318524]
Temporal grounding is the task of locating a specific segment from an untrimmed video according to a query sentence.
We introduce a new Compositional Temporal Grounding task and construct two new dataset splits.
We argue that the inherent structured semantics inside the videos and language is the crucial factor to achieve compositional generalization.
arXiv Detail & Related papers (2023-01-22T08:02:23Z) - Textual Entailment Recognition with Semantic Features from Empirical
Text Representation [60.31047947815282]
A text entails a hypothesis if and only if the true value of the hypothesis follows the text.
In this paper, we propose a novel approach to identifying the textual entailment relationship between text and hypothesis.
We employ an element-wise Manhattan distance vector-based feature that can identify the semantic entailment relationship between the text-hypothesis pair.
arXiv Detail & Related papers (2022-10-18T10:03:51Z) - Distilling Relation Embeddings from Pre-trained Language Models [35.718167335989854]
We show that it is possible to distill relation embeddings from pre-trained language models.
We encode word pairs using a (manually or automatically generated) prompt, and we fine-tune the language model.
The resulting relation embeddings are highly competitive on analogy (unsupervised) and relation classification (supervised) benchmarks.
arXiv Detail & Related papers (2021-09-21T15:05:27Z) - EDS-MEMBED: Multi-sense embeddings based on enhanced distributional
semantic structures via a graph walk over word senses [0.0]
We leverage the rich semantic structures in WordNet to enhance the quality of multi-sense embeddings.
We derive new distributional semantic similarity measures for M-SE from prior ones.
We report evaluation results on 11 benchmark datasets involving WSD and Word Similarity tasks.
arXiv Detail & Related papers (2021-02-27T14:36:55Z) - Using Holographically Compressed Embeddings in Question Answering [0.0]
This research employs holographic compression of pre-trained embeddings to represent a token, its part-of-speech, and named entity type.
The implementation, in a modified question answering recurrent deep learning network, shows that semantic relationships are preserved, and yields strong performance.
arXiv Detail & Related papers (2020-07-14T18:29:49Z) - On the Learnability of Concepts: With Applications to Comparing Word
Embedding Algorithms [0.0]
We introduce the notion of "concept" as a list of words that have shared semantic content.
We first use this notion to measure the learnability of concepts on pretrained word embeddings.
We then develop a statistical analysis of concept learnability, based on hypothesis testing and ROC curves, in order to compare the relative merits of various embedding algorithms.
arXiv Detail & Related papers (2020-06-17T14:25:36Z) - On Vocabulary Reliance in Scene Text Recognition [79.21737876442253]
Methods perform well on images with words within vocabulary but generalize poorly to images with words outside vocabulary.
We call this phenomenon "vocabulary reliance"
We propose a simple yet effective mutual learning strategy to allow models of two families to learn collaboratively.
arXiv Detail & Related papers (2020-05-08T11:16:58Z) - Comparative Analysis of Word Embeddings for Capturing Word Similarities [0.0]
Distributed language representation has become the most widely used technique for language representation in various natural language processing tasks.
Most of the natural language processing models that are based on deep learning techniques use already pre-trained distributed word representations, commonly called word embeddings.
selecting the appropriate word embeddings is a perplexing task since the projected embedding space is not intuitive to humans.
arXiv Detail & Related papers (2020-05-08T01:16:03Z) - Lexical Sememe Prediction using Dictionary Definitions by Capturing
Local Semantic Correspondence [94.79912471702782]
Sememes, defined as the minimum semantic units of human languages, have been proven useful in many NLP tasks.
We propose a Sememe Correspondence Pooling (SCorP) model, which is able to capture this kind of matching to predict sememes.
We evaluate our model and baseline methods on a famous sememe KB HowNet and find that our model achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-01-16T17:30:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.