Revisiting Entropy Rate Constancy in Text
- URL: http://arxiv.org/abs/2305.12084v2
- Date: Wed, 18 Oct 2023 01:02:56 GMT
- Title: Revisiting Entropy Rate Constancy in Text
- Authors: Vivek Verma, Nicholas Tomlin, Dan Klein
- Abstract summary: The uniform information density hypothesis states that humans tend to distribute information roughly evenly across an utterance or discourse.
We re-evaluate the claims of Genzel & Charniak (2002) with neural language models, failing to find clear evidence in support of entropy rate constancy.
- Score: 43.928576088761844
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The uniform information density (UID) hypothesis states that humans tend to
distribute information roughly evenly across an utterance or discourse. Early
evidence in support of the UID hypothesis came from Genzel & Charniak (2002),
which proposed an entropy rate constancy principle based on the probability of
English text under n-gram language models. We re-evaluate the claims of Genzel
& Charniak (2002) with neural language models, failing to find clear evidence
in support of entropy rate constancy. We conduct a range of experiments across
datasets, model sizes, and languages and discuss implications for the uniform
information density hypothesis and linguistic theories of efficient
communication more broadly.
Related papers
- Surprise! Uniform Information Density Isn't the Whole Story: Predicting Surprisal Contours in Long-form Discourse [54.08750245737734]
We propose that speakers modulate information rate based on location within a hierarchically-structured model of discourse.
We find that hierarchical predictors are significant predictors of a discourse's information contour and that deeply nested hierarchical predictors are more predictive than shallow ones.
arXiv Detail & Related papers (2024-10-21T14:42:37Z) - A generative framework to bridge data-driven models and scientific theories in language neuroscience [84.76462599023802]
We present generative explanation-mediated validation, a framework for generating concise explanations of language selectivity in the brain.
We show that explanatory accuracy is closely related to the predictive power and stability of the underlying statistical models.
arXiv Detail & Related papers (2024-10-01T15:57:48Z) - That's Optional: A Contemporary Exploration of "that" Omission in English Subordinate Clauses [2.1781981800541805]
The Uniform Information Density hypothesis posits that speakers optimize the communicative properties of their utterances by avoiding spikes in information.
This paper investigates the impact of UID principles on syntactic reduction, specifically focusing on the optional omission of the connector "that" in English subordinate clauses.
arXiv Detail & Related papers (2024-05-31T14:23:30Z) - Estimating the Entropy of Linguistic Distributions [75.20045001387685]
We study the empirical effectiveness of different entropy estimators for linguistic distributions.
We find evidence that the reported effect size is over-estimated due to over-reliance on poor entropy estimators.
arXiv Detail & Related papers (2022-04-04T13:36:46Z) - On the probability-quality paradox in language generation [76.69397802617064]
We analyze language generation through an information-theoretic lens.
We posit that human-like language should contain an amount of information close to the entropy of the distribution over natural strings.
arXiv Detail & Related papers (2022-03-31T17:43:53Z) - Revisiting the Uniform Information Density Hypothesis [44.277066511088634]
We investigate the uniform information density (UID) hypothesis using reading time and acceptability data.
For acceptability judgments, we find clearer evidence that non-uniformity in information density is predictive of lower acceptability.
arXiv Detail & Related papers (2021-09-23T20:41:47Z) - A Cognitive Regularizer for Language Modeling [36.256053903862956]
We augment the canonical MLE objective for training language models by encoding UID as regularization.
We find that using UID regularization consistently improves perplexity in language models.
We also find that UID-regularized language models are higher-entropy and produce text that is longer and more lexically diverse.
arXiv Detail & Related papers (2021-05-15T05:37:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.