Semantic Holism and Word Representations in Artificial Neural Networks
- URL: http://arxiv.org/abs/2003.05522v1
- Date: Wed, 11 Mar 2020 21:04:49 GMT
- Title: Semantic Holism and Word Representations in Artificial Neural Networks
- Authors: Tom\'a\v{s} Musil
- Abstract summary: We show that word representations from the Skip-gram variant of the word2vec model exhibit interesting semantic properties.
This is usually explained by referring to the general distributional hypothesis.
We propose a more specific approach based on Frege's holistic and functional approach to meaning.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Artificial neural networks are a state-of-the-art solution for many problems
in natural language processing. What can we learn about language and meaning
from the way artificial neural networks represent it? Word representations
obtained from the Skip-gram variant of the word2vec model exhibit interesting
semantic properties. This is usually explained by referring to the general
distributional hypothesis, which states that the meaning of the word is given
by the contexts where it occurs. We propose a more specific approach based on
Frege's holistic and functional approach to meaning. Taking Tugendhat's formal
reinterpretation of Frege's work as a starting point, we demonstrate that it is
analogical to the process of training the Skip-gram model and offers a possible
explanation of its semantic properties.
Related papers
- Neural Semantic Parsing with Extremely Rich Symbolic Meaning Representations [7.774674200374255]
We introduce a novel compositional symbolic representation for concepts based on their position in the taxonomical hierarchy.
This representation provides richer semantic information and enhances interpretability.
Our experimental findings demonstrate that the taxonomical model, trained on much richer and complex meaning representations, is slightly subordinate in performance to the traditional model using the standard metrics for evaluation, but outperforms it when dealing with out-of-vocabulary concepts.
arXiv Detail & Related papers (2024-04-19T08:06:01Z) - Grounded learning for compositional vector semantics [1.4344589271451351]
This work proposes a way for compositional distributional semantics to be implemented within a spiking neural network architecture.
We also describe a means of training word representations using labelled images.
arXiv Detail & Related papers (2024-01-10T22:12:34Z) - Probabilistic Transformer: A Probabilistic Dependency Model for
Contextual Word Representation [52.270712965271656]
We propose a new model of contextual word representation, not from a neural perspective, but from a purely syntactic and probabilistic perspective.
We find that the graph of our model resembles transformers, with correspondences between dependencies and self-attention.
Experiments show that our model performs competitively to transformers on small to medium sized datasets.
arXiv Detail & Related papers (2023-11-26T06:56:02Z) - Agentivit\`a e telicit\`a in GilBERTo: implicazioni cognitive [77.71680953280436]
The goal of this study is to investigate whether a Transformer-based neural language model infers lexical semantics.
The semantic properties considered are telicity (also combined with definiteness) and agentivity.
arXiv Detail & Related papers (2023-07-06T10:52:22Z) - Seeing in Words: Learning to Classify through Language Bottlenecks [59.97827889540685]
Humans can explain their predictions using succinct and intuitive descriptions.
We show that a vision model whose feature representations are text can effectively classify ImageNet images.
arXiv Detail & Related papers (2023-06-29T00:24:42Z) - Deep Learning Models to Study Sentence Comprehension in the Human Brain [0.1503974529275767]
Recent artificial neural networks that process natural language achieve unprecedented performance in tasks requiring sentence-level understanding.
We review works that compare these artificial language models with human brain activity and we assess the extent to which this approach has improved our understanding of the neural processes involved in natural language comprehension.
arXiv Detail & Related papers (2023-01-16T10:31:25Z) - Latent Topology Induction for Understanding Contextualized
Representations [84.7918739062235]
We study the representation space of contextualized embeddings and gain insight into the hidden topology of large language models.
We show there exists a network of latent states that summarize linguistic properties of contextualized representations.
arXiv Detail & Related papers (2022-06-03T11:22:48Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Explaining Deep Neural Networks [12.100913944042972]
In various domains, such as healthcare, finance, or law, it is critical to know the reasons behind a decision made by an artificial intelligence system.
This thesis investigates two major directions for explaining deep neural networks.
arXiv Detail & Related papers (2020-10-04T07:23:13Z) - The Representation Theory of Neural Networks [7.724617675868718]
We show that neural networks can be represented via the mathematical theory of quiver representations.
We show that network quivers gently adapt to common neural network concepts.
We also provide a quiver representation model to understand how a neural network creates representations from the data.
arXiv Detail & Related papers (2020-07-23T19:02:14Z) - Compositional Explanations of Neurons [52.71742655312625]
We describe a procedure for explaining neurons in deep representations by identifying compositional logical concepts.
We use this procedure to answer several questions on interpretability in models for vision and natural language processing.
arXiv Detail & Related papers (2020-06-24T20:37:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.