What's in a Scientific Name?
- URL: http://arxiv.org/abs/2106.14610v1
- Date: Mon, 31 May 2021 22:06:20 GMT
- Title: What's in a Scientific Name?
- Authors: Henrique Ferraz de Arruda, Luciano da Fontoura Costa
- Abstract summary: The study reported here takes into account the words "prediction", "model", "optimization", "complex", "entropy", "random", "deterministic", "pattern", and "Database"
Several of the words were observed to have markedly distinct associations in different areas. Biology was found to be related to computer science, sharing associations with databases.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To a good extent, words can be understood as corresponding to patterns or
categories that appeared in order to represent concepts and structures that are
particularly important or useful in a given time and space. Words are
characterized by not being completely general nor specific, in the sense that
the same word can be instantiated or related to several different contexts,
depending on specific situations. Indeed, the way in which words are
instantiated and associated represents a particularly interesting aspect that
can substantially help to better understand the context in which they are
employed. Scientific words are no exception to that. In the present work, we
approach the associations between a set of particularly relevant words in the
sense of being not only frequently used in several areas, but also representing
concepts that are currently related to some of the main standing challenges in
science. More specifically, the study reported here takes into account the
words "prediction", "model", "optimization", "complex", "entropy", "random",
"deterministic", "pattern", and "database". In order to complement the
analysis, we also obtain a network representing the relationship between the
adopted areas. Many interesting results were found. First and foremost, several
of the words were observed to have markedly distinct associations in different
areas. Biology was found to be related to computer science, sharing
associations with databases. Furthermore, for most of the cases, the words
"complex", "model", and "prediction" were observed to have several strong
associations.
Related papers
- Measuring Meaning Composition in the Human Brain with Composition Scores from Large Language Models [53.840982361119565]
The Composition Score is a novel model-based metric designed to quantify the degree of meaning composition during sentence comprehension.
Experimental findings show that this metric correlates with brain clusters associated with word frequency, structural processing, and general sensitivity to words.
arXiv Detail & Related papers (2024-03-07T08:44:42Z) - How Do In-Context Examples Affect Compositional Generalization? [86.57079616209474]
In this paper, we present CoFe, a test suite to investigate in-context compositional generalization.
We find that the compositional generalization performance can be easily affected by the selection of in-context examples.
Our systematic experiments indicate that in-context examples should be structurally similar to the test case, diverse from each other, and individually simple.
arXiv Detail & Related papers (2023-05-08T16:32:18Z) - Neighboring Words Affect Human Interpretation of Saliency Explanations [65.29015910991261]
Word-level saliency explanations are often used to communicate feature-attribution in text-based models.
Recent studies found that superficial factors such as word length can distort human interpretation of the communicated saliency scores.
We investigate how the marking of a word's neighboring words affect the explainee's perception of the word's importance in the context of a saliency explanation.
arXiv Detail & Related papers (2023-05-04T09:50:25Z) - Quantifying the Roles of Visual, Linguistic, and Visual-Linguistic
Complexity in Verb Acquisition [8.183763443800348]
We employ visual and linguistic representations of words sourced from pre-trained artificial neural networks.
We find that the representation of verbs is generally more variable and less discriminable within domain than the representation of nouns.
Visual variability is the strongest factor that internally drives verb learning, followed by visual-linguistic alignment and linguistic variability.
arXiv Detail & Related papers (2023-04-05T15:08:21Z) - Evaluation of Word Embeddings for the Social Sciences [0.0]
We describe the creation and evaluation of word embedding models based on 37,604 social science research papers.
We found that the created domain-specific model covers a large part of social science concepts.
Across all relation types, we found a more extensive coverage of semantic relationships.
arXiv Detail & Related papers (2023-02-13T08:23:03Z) - Words as Gatekeepers: Measuring Discipline-specific Terms and Meanings
in Scholarly Publications [13.443742417911352]
We develop and validate an interpretable approach for measuring scholarly jargon from text.
We use word sense induction to identify words that are widespread but overloaded with different meanings across fields.
We show that word senses provide a complementary, yet unique view of jargon alongside word types.
arXiv Detail & Related papers (2022-12-19T18:01:06Z) - An Informational Space Based Semantic Analysis for Scientific Texts [62.997667081978825]
This paper introduces computational methods for semantic analysis and the quantifying the meaning of short scientific texts.
The representation of scientific-specific meaning is standardised by replacing the situation representations, rather than psychological properties.
The research in this paper conducts the base for the geometric representation of the meaning of texts.
arXiv Detail & Related papers (2022-05-31T11:19:32Z) - Towards a Theoretical Understanding of Word and Relation Representation [8.020742121274418]
Representing words by vectors, or embeddings, enables computational reasoning.
We focus on word embeddings learned from text corpora and knowledge graphs.
arXiv Detail & Related papers (2022-02-01T15:34:58Z) - Compositional Processing Emerges in Neural Networks Solving Math
Problems [100.80518350845668]
Recent progress in artificial neural networks has shown that when large models are trained on enough linguistic data, grammatical structure emerges in their representations.
We extend this work to the domain of mathematical reasoning, where it is possible to formulate precise hypotheses about how meanings should be composed.
Our work shows that neural networks are not only able to infer something about the structured relationships implicit in their training data, but can also deploy this knowledge to guide the composition of individual meanings into composite wholes.
arXiv Detail & Related papers (2021-05-19T07:24:42Z) - Topology of Word Embeddings: Singularities Reflect Polysemy [68.8204255655161]
We introduce a topological measure of polysemy based on persistent homology that correlates well with the actual number of meanings of a word.
We propose a simple, topologically motivated solution to the SemEval-2010 task on Word Sense Induction & Disambiguation.
arXiv Detail & Related papers (2020-11-18T17:21:51Z) - Contextualized Word Embeddings Encode Aspects of Human-Like Word Sense
Knowledge [0.0]
We investigate whether recent advances in NLP, specifically contextualized word embeddings, capture human-like distinctions between English word senses.
We find that participants' judgments of the relatedness between senses are correlated with distances between senses in the BERT embedding space.
Our findings point towards the potential utility of continuous-space representations of sense meanings.
arXiv Detail & Related papers (2020-10-25T07:56:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.