Quantifying the Roles of Visual, Linguistic, and Visual-Linguistic
Complexity in Verb Acquisition
- URL: http://arxiv.org/abs/2304.02492v1
- Date: Wed, 5 Apr 2023 15:08:21 GMT
- Title: Quantifying the Roles of Visual, Linguistic, and Visual-Linguistic
Complexity in Verb Acquisition
- Authors: Yuchen Zhou, Michael J. Tarr, Daniel Yurovsky
- Abstract summary: We employ visual and linguistic representations of words sourced from pre-trained artificial neural networks.
We find that the representation of verbs is generally more variable and less discriminable within domain than the representation of nouns.
Visual variability is the strongest factor that internally drives verb learning, followed by visual-linguistic alignment and linguistic variability.
- Score: 8.183763443800348
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Children typically learn the meanings of nouns earlier than the meanings of
verbs. However, it is unclear whether this asymmetry is a result of complexity
in the visual structure of categories in the world to which language refers,
the structure of language itself, or the interplay between the two sources of
information. We quantitatively test these three hypotheses regarding early verb
learning by employing visual and linguistic representations of words sourced
from large-scale pre-trained artificial neural networks. Examining the
structure of both visual and linguistic embedding spaces, we find, first, that
the representation of verbs is generally more variable and less discriminable
within domain than the representation of nouns. Second, we find that if only
one learning instance per category is available, visual and linguistic
representations are less well aligned in the verb system than in the noun
system. However, in parallel with the course of human language development, if
multiple learning instances per category are available, visual and linguistic
representations become almost as well aligned in the verb system as in the noun
system. Third, we compare the relative contributions of factors that may
predict learning difficulty for individual words. A regression analysis reveals
that visual variability is the strongest factor that internally drives verb
learning, followed by visual-linguistic alignment and linguistic variability.
Based on these results, we conclude that verb acquisition is influenced by all
three sources of complexity, but that the variability of visual structure poses
the most significant challenge for verb learning.
Related papers
- Reframing linguistic bootstrapping as joint inference using visually-grounded grammar induction models [31.006803764376475]
Semantic and syntactic bootstrapping posit that children use their prior knowledge of one linguistic domain, say syntactic relations, to help later acquire another, such as the meanings of new words.
Here, we argue that they are instead both contingent on a more general learning strategy for language acquisition: joint learning.
Using a series of neural visually-grounded grammar induction models, we demonstrate that both syntactic and semantic bootstrapping effects are strongest when syntax and semantics are learnt simultaneously.
arXiv Detail & Related papers (2024-06-17T18:01:06Z) - Word class representations spontaneously emerge in a deep neural network
trained on next word prediction [7.240611820374677]
How do humans learn language, and can the first language be learned at all?
These fundamental questions are still hotly debated.
In particular, we train an artificial deep neural network on predicting the next word.
We find that the internal representations of nine-word input sequences cluster according to the word class of the tenth word to be predicted as output.
arXiv Detail & Related papers (2023-02-15T11:02:50Z) - A Linguistic Investigation of Machine Learning based Contradiction
Detection Models: An Empirical Analysis and Future Perspectives [0.34998703934432673]
We analyze two Natural Language Inference data sets with respect to their linguistic features.
The goal is to identify those syntactic and semantic properties that are particularly hard to comprehend for a machine learning model.
arXiv Detail & Related papers (2022-10-19T10:06:03Z) - Transparency Helps Reveal When Language Models Learn Meaning [71.96920839263457]
Our systematic experiments with synthetic data reveal that, with languages where all expressions have context-independent denotations, both autoregressive and masked language models learn to emulate semantic relations between expressions.
Turning to natural language, our experiments with a specific phenomenon -- referential opacity -- add to the growing body of evidence that current language models do not well-represent natural language semantics.
arXiv Detail & Related papers (2022-10-14T02:35:19Z) - Exploring the Representation of Word Meanings in Context: A Case Study
on Homonymy and Synonymy [0.0]
We assess the ability of both static and contextualized models to adequately represent different lexical-semantic relations.
Experiments are performed in Galician, Portuguese, English, and Spanish.
arXiv Detail & Related papers (2021-06-25T10:54:23Z) - AM2iCo: Evaluating Word Meaning in Context across Low-ResourceLanguages
with Adversarial Examples [51.048234591165155]
We present AM2iCo, Adversarial and Multilingual Meaning in Context.
It aims to faithfully assess the ability of state-of-the-art (SotA) representation models to understand the identity of word meaning in cross-lingual contexts.
Results reveal that current SotA pretrained encoders substantially lag behind human performance.
arXiv Detail & Related papers (2021-04-17T20:23:45Z) - VLGrammar: Grounded Grammar Induction of Vision and Language [86.88273769411428]
We study grounded grammar induction of vision and language in a joint learning framework.
We present VLGrammar, a method that uses compound probabilistic context-free grammars (compound PCFGs) to induce the language grammar and the image grammar simultaneously.
arXiv Detail & Related papers (2021-03-24T04:05:08Z) - Word Frequency Does Not Predict Grammatical Knowledge in Language Models [2.1984302611206537]
We investigate whether there are systematic sources of variation in the language models' accuracy.
We find that certain nouns are systematically understood better than others, an effect which is robust across grammatical tasks and different language models.
We find that a novel noun's grammatical properties can be few-shot learned from various types of training data.
arXiv Detail & Related papers (2020-10-26T19:51:36Z) - Investigating Cross-Linguistic Adjective Ordering Tendencies with a
Latent-Variable Model [66.84264870118723]
We present the first purely corpus-driven model of multi-lingual adjective ordering in the form of a latent-variable model.
We provide strong converging evidence for the existence of universal, cross-linguistic, hierarchical adjective ordering tendencies.
arXiv Detail & Related papers (2020-10-09T18:27:55Z) - Predicting Declension Class from Form and Meaning [70.65971611552871]
Class membership is far from deterministic, but the phonological form of a noun and/or its meaning can often provide imperfect clues.
We operationalize this by measuring how much information, in bits, we can glean about declension class from knowing the form and/or meaning of nouns.
We find for two Indo-European languages (Czech and German) that form and meaning respectively share significant amounts of information with class.
arXiv Detail & Related papers (2020-05-01T21:48:48Z) - Where New Words Are Born: Distributional Semantic Analysis of Neologisms
and Their Semantic Neighborhoods [51.34667808471513]
We investigate the importance of two factors, semantic sparsity and frequency growth rates of semantic neighbors, formalized in the distributional semantics paradigm.
We show that both factors are predictive word emergence although we find more support for the latter hypothesis.
arXiv Detail & Related papers (2020-01-21T19:09:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.