Prosody leaks into the memories of words
- URL: http://arxiv.org/abs/2005.14716v2
- Date: Tue, 29 Dec 2020 16:09:11 GMT
- Title: Prosody leaks into the memories of words
- Authors: Kevin Tang, Jason A. Shaw
- Abstract summary: The average predictability (aka informativity) of a word in context has been shown to condition word duration.
This study extends past work in two directions; it investigated informativity effects in another large language, Mandarin Chinese.
Results indicated that words with low informativity have shorter durations, replicating the effect found in English.
- Score: 2.309770674164469
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The average predictability (aka informativity) of a word in context has been
shown to condition word duration (Seyfarth, 2014). All else being equal, words
that tend to occur in more predictable environments are shorter than words that
tend to occur in less predictable environments. One account of the
informativity effect on duration is that the acoustic details of probabilistic
reduction are stored as part of a word's mental representation. Other research
has argued that predictability effects are tied to prosodic structure in
integral ways. With the aim of assessing a potential prosodic basis for
informativity effects in speech production, this study extends past work in two
directions; it investigated informativity effects in another large language,
Mandarin Chinese, and broadened the study beyond word duration to additional
acoustic dimensions, pitch and intensity, known to index prosodic prominence.
The acoustic information of content words was extracted from a large telephone
conversation speech corpus with over 400,000 tokens and 6,000 word types spoken
by 1,655 individuals and analyzed for the effect of informativity using
frequency statistics estimated from a 431 million word subtitle corpus. Results
indicated that words with low informativity have shorter durations, replicating
the effect found in English. In addition, informativity had significant effects
on maximum pitch and intensity, two phonetic dimensions related to prosodic
prominence. Extending this interpretation, these results suggest that
predictability is closely linked to prosodic prominence, and that the lexical
representation of a word includes phonetic details associated with its average
prosodic prominence in discourse. In other words, the lexicon absorbs prosodic
influences on speech production.
Related papers
- Surprise! Uniform Information Density Isn't the Whole Story: Predicting Surprisal Contours in Long-form Discourse [54.08750245737734]
We propose that speakers modulate information rate based on location within a hierarchically-structured model of discourse.
We find that hierarchical predictors are significant predictors of a discourse's information contour and that deeply nested hierarchical predictors are more predictive than shallow ones.
arXiv Detail & Related papers (2024-10-21T14:42:37Z) - Quantifying the redundancy between prosody and text [67.07817268372743]
We use large language models to estimate how much information is redundant between prosody and the words themselves.
We find a high degree of redundancy between the information carried by the words and prosodic information across several prosodic features.
Still, we observe that prosodic features can not be fully predicted from text, suggesting that prosody carries information above and beyond the words.
arXiv Detail & Related papers (2023-11-28T21:15:24Z) - Improving Mandarin Prosodic Structure Prediction with Multi-level
Contextual Information [68.89000132126536]
This work proposes to use inter-utterance linguistic information to improve the performance of prosodic structure prediction (PSP)
Our method achieves better F1 scores in predicting prosodic word (PW), prosodic phrase (PPH) and intonational phrase (IPH)
arXiv Detail & Related papers (2023-08-31T09:19:15Z) - Investigating the Utility of Surprisal from Large Language Models for
Speech Synthesis Prosody [4.081433571732691]
This paper investigates the use of word surprisal, a measure of the predictability of a word in a given context, as a feature to aid speech prosody synthesis.
We conduct experiments using a large corpus of English text and large language models (LLMs) of varying sizes.
We find that word surprisal and word prominence are moderately correlated, suggesting that they capture related but distinct aspects of language use.
arXiv Detail & Related papers (2023-06-16T12:49:44Z) - Neighboring Words Affect Human Interpretation of Saliency Explanations [65.29015910991261]
Word-level saliency explanations are often used to communicate feature-attribution in text-based models.
Recent studies found that superficial factors such as word length can distort human interpretation of the communicated saliency scores.
We investigate how the marking of a word's neighboring words affect the explainee's perception of the word's importance in the context of a saliency explanation.
arXiv Detail & Related papers (2023-05-04T09:50:25Z) - Discourse Context Predictability Effects in Hindi Word Order [14.88833412862455]
We investigate how the words and syntactic structures in a sentence influence the word order of the following sentences.
We use a number of discourse-based features and cognitive features to make its predictions, including dependency length, surprisal, and information status.
We find that information status and LSTM-based discourse predictability influence word order choices, especially for non-canonical object-fronted orders.
arXiv Detail & Related papers (2022-10-25T11:53:01Z) - Unsupervised Multimodal Word Discovery based on Double Articulation
Analysis with Co-occurrence cues [7.332652485849632]
Human infants acquire their verbal lexicon with minimal prior knowledge of language.
This study proposes a novel fully unsupervised learning method for discovering speech units.
The proposed method can acquire words and phonemes from speech signals using unsupervised learning.
arXiv Detail & Related papers (2022-01-18T07:31:59Z) - Disambiguatory Signals are Stronger in Word-initial Positions [48.18148856974974]
We point out the confounds in existing methods for comparing the informativeness of segments early in the word versus later in the word.
We find evidence across hundreds of languages that indeed there is a cross-linguistic tendency to front-load information in words.
arXiv Detail & Related papers (2021-02-03T18:19:16Z) - Speakers Fill Lexical Semantic Gaps with Context [65.08205006886591]
We operationalise the lexical ambiguity of a word as the entropy of meanings it can take.
We find significant correlations between our estimate of ambiguity and the number of synonyms a word has in WordNet.
This suggests that, in the presence of ambiguity, speakers compensate by making contexts more informative.
arXiv Detail & Related papers (2020-10-05T17:19:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.