Inter-Sense: An Investigation of Sensory Blending in Fiction
- URL: http://arxiv.org/abs/2110.09710v1
- Date: Tue, 19 Oct 2021 03:25:26 GMT
- Title: Inter-Sense: An Investigation of Sensory Blending in Fiction
- Authors: Roxana Girju and Charlotte Lambert
- Abstract summary: This study reports on the semantic organization of English sensory descriptors of sight, hearing, touch, taste, and smell in a large corpus of over 8,000 fiction books.
We introduce a large-scale text data-driven approach based on distributional-semantic word embeddings to identify and extract these descriptors.
The findings are relevant for research on concept acquisition and representation, as well as for applications that can benefit from a better understanding of perceptual spaces of sensory experiences.
- Score: 0.40611352512781856
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This study reports on the semantic organization of English sensory
descriptors of the five basic senses of sight, hearing, touch, taste, and smell
in a large corpus of over 8,000 fiction books. We introduce a large-scale text
data-driven approach based on distributional-semantic word embeddings to
identify and extract these descriptors as well as analyze their mixing
interconnections in the resulting conceptual and sensory space. The findings
are relevant for research on concept acquisition and representation, as well as
for applications that can benefit from a better understanding of perceptual
spaces of sensory experiences, in fiction, in particular, and in language in
general.
Related papers
- To Word Senses and Beyond: Inducing Concepts with Contextualized Language Models [0.9176056742068812]
Polysemy and synonymy are crucial facets of lexical ambiguity.
In this paper, we introduce Concept Induction, the unsupervised task of learning a soft clustering among words.
We propose a bi-level approach to Concept Induction that leverages both a local lemma-centric view and a global cross-lexicon perspective.
arXiv Detail & Related papers (2024-06-28T17:07:06Z) - A Survey on Lexical Ambiguity Detection and Word Sense Disambiguation [0.0]
This paper explores techniques that focus on understanding and resolving ambiguity in language within the field of natural language processing (NLP)
It outlines diverse approaches ranging from deep learning techniques to leveraging lexical resources and knowledge graphs like WordNet.
The research identifies persistent challenges in the field, such as the scarcity of sense annotated corpora and the complexity of informal clinical texts.
arXiv Detail & Related papers (2024-03-24T12:58:48Z) - Measuring Meaning Composition in the Human Brain with Composition Scores from Large Language Models [53.840982361119565]
The Composition Score is a novel model-based metric designed to quantify the degree of meaning composition during sentence comprehension.
Experimental findings show that this metric correlates with brain clusters associated with word frequency, structural processing, and general sensitivity to words.
arXiv Detail & Related papers (2024-03-07T08:44:42Z) - Semantic Specialization for Knowledge-based Word Sense Disambiguation [12.573927420408365]
A promising approach for knowledge-based Word Sense Disambiguation (WSD) is to select the sense whose contextualized embeddings are closest to those computed for a target word in a given sentence.
We propose a semantic specialization for WSD where contextualized embeddings are adapted to the WSD task using solely lexical knowledge.
arXiv Detail & Related papers (2023-04-22T07:40:23Z) - An Informational Space Based Semantic Analysis for Scientific Texts [62.997667081978825]
This paper introduces computational methods for semantic analysis and the quantifying the meaning of short scientific texts.
The representation of scientific-specific meaning is standardised by replacing the situation representations, rather than psychological properties.
The research in this paper conducts the base for the geometric representation of the meaning of texts.
arXiv Detail & Related papers (2022-05-31T11:19:32Z) - Exploring the Sensory Spaces of English Perceptual Verbs in Natural
Language Data [0.40611352512781856]
We focus on the most frequent perception verbs of English analyzed from an and Agentive vs. Experiential distinction.
In this study we report on a data-driven approach based on distributional-semantic word embeddings and clustering models.
arXiv Detail & Related papers (2021-10-19T03:58:44Z) - Large Scale Substitution-based Word Sense Induction [48.49573297876054]
We present a word-sense induction method based on pre-trained masked language models (MLMs), which can cheaply scale to large vocabularies and large corpora.
The result is a corpus which is sense-tagged according to a corpus-derived sense inventory and where each sense is associated with indicative words.
Evaluation on English Wikipedia that was sense-tagged using our method shows that both the induced senses, and the per-instance sense assignment, are of high quality even compared to WSD methods, such as Babelfy.
arXiv Detail & Related papers (2021-10-14T19:40:37Z) - Accessible Visualization via Natural Language Descriptions: A Four-Level
Model of Semantic Content [6.434361163743876]
We introduce a conceptual model for the semantic content conveyed by natural language descriptions of visualizations.
We conduct a mixed-methods evaluation with 30 blind and 90 sighted readers, and find that these reader groups differ significantly on which semantic content they rank as most useful.
arXiv Detail & Related papers (2021-10-08T23:37:25Z) - YouRefIt: Embodied Reference Understanding with Language and Gesture [95.93218436323481]
We study the understanding of embodied reference.
One agent uses both language and gesture to refer to an object to another agent in a shared physical environment.
Crowd-sourced YouRefIt dataset contains 4,195 unique reference clips in 432 indoor scenes.
arXiv Detail & Related papers (2021-09-08T03:27:32Z) - Constellation: Learning relational abstractions over objects for
compositional imagination [64.99658940906917]
We introduce Constellation, a network that learns relational abstractions of static visual scenes.
This work is a first step in the explicit representation of visual relationships and using them for complex cognitive procedures.
arXiv Detail & Related papers (2021-07-23T11:59:40Z) - Clinical Named Entity Recognition using Contextualized Token
Representations [49.036805795072645]
This paper introduces the technique of contextualized word embedding to better capture the semantic meaning of each word based on its context.
We pre-train two deep contextualized language models, Clinical Embeddings from Language Model (C-ELMo) and Clinical Contextual String Embeddings (C-Flair)
Explicit experiments show that our models gain dramatic improvements compared to both static word embeddings and domain-generic language models.
arXiv Detail & Related papers (2021-06-23T18:12:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.