Lexical semantic change for Ancient Greek and Latin
- URL: http://arxiv.org/abs/2101.09069v1
- Date: Fri, 22 Jan 2021 12:04:08 GMT
- Title: Lexical semantic change for Ancient Greek and Latin
- Authors: Valerio Perrone and Simon Hengchen and Marco Palma and Alessandro
Vatri and Jim Q. Smith and Barbara McGillivray
- Abstract summary: Associating a word's correct meaning in its historical context is a central challenge in diachronic research.
We build on a recent computational approach to semantic change based on a dynamic Bayesian mixture model.
We provide a systematic comparison of dynamic Bayesian mixture models for semantic change with state-of-the-art embedding-based models.
- Score: 61.69697586178796
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Change and its precondition, variation, are inherent in languages. Over time,
new words enter the lexicon, others become obsolete, and existing words acquire
new senses. Associating a word's correct meaning in its historical context is a
central challenge in diachronic research. Historical corpora of classical
languages, such as Ancient Greek and Latin, typically come with rich metadata,
and existing models are limited by their inability to exploit contextual
information beyond the document timestamp. While embedding-based methods
feature among the current state of the art systems, they are lacking in the
interpretative power. In contrast, Bayesian models provide explicit and
interpretable representations of semantic change phenomena. In this chapter we
build on GASC, a recent computational approach to semantic change based on a
dynamic Bayesian mixture model. In this model, the evolution of word senses
over time is based not only on distributional information of lexical nature,
but also on text genres. We provide a systematic comparison of dynamic Bayesian
mixture models for semantic change with state-of-the-art embedding-based
models. On top of providing a full description of meaning change over time, we
show that Bayesian mixture models are highly competitive approaches to detect
binary semantic change in both Ancient Greek and Latin.
Related papers
- Using Synchronic Definitions and Semantic Relations to Classify Semantic Change Types [1.3436368800886478]
We present a model that leverages information both from synchronic lexical relations and definitions of word meanings.
Specifically, we use synset definitions and hierarchy information from WordNet and test it on a digitized version of Blank's (1997) dataset of semantic change types.
arXiv Detail & Related papers (2024-06-05T16:52:21Z) - Analyzing Semantic Change through Lexical Replacements [2.509907053583601]
We study the effect of unexpected contexts introduced by textitlexical replacements
We propose a textitreplacement schema where a target word is substituted with lexical replacements of varying relatedness.
We are the first to evaluate the use of LLaMa for semantic change detection.
arXiv Detail & Related papers (2024-04-29T10:20:41Z) - Variational Cross-Graph Reasoning and Adaptive Structured Semantics
Learning for Compositional Temporal Grounding [143.5927158318524]
Temporal grounding is the task of locating a specific segment from an untrimmed video according to a query sentence.
We introduce a new Compositional Temporal Grounding task and construct two new dataset splits.
We argue that the inherent structured semantics inside the videos and language is the crucial factor to achieve compositional generalization.
arXiv Detail & Related papers (2023-01-22T08:02:23Z) - Transparency Helps Reveal When Language Models Learn Meaning [71.96920839263457]
Our systematic experiments with synthetic data reveal that, with languages where all expressions have context-independent denotations, both autoregressive and masked language models learn to emulate semantic relations between expressions.
Turning to natural language, our experiments with a specific phenomenon -- referential opacity -- add to the growing body of evidence that current language models do not well-represent natural language semantics.
arXiv Detail & Related papers (2022-10-14T02:35:19Z) - Contextualized language models for semantic change detection: lessons
learned [4.436724861363513]
We present a qualitative analysis of the outputs of contextualized embedding-based methods for detecting diachronic semantic change.
Our findings show that contextualized methods can often predict high change scores for words which are not undergoing any real diachronic semantic shift.
Our conclusion is that pre-trained contextualized language models are prone to confound changes in lexicographic senses and changes in contextual variance.
arXiv Detail & Related papers (2022-08-31T23:35:24Z) - Compositional Temporal Grounding with Structured Variational Cross-Graph
Correspondence Learning [92.07643510310766]
Temporal grounding in videos aims to localize one target video segment that semantically corresponds to a given query sentence.
We introduce a new Compositional Temporal Grounding task and construct two new dataset splits.
We empirically find that they fail to generalize to queries with novel combinations of seen words.
We propose a variational cross-graph reasoning framework that explicitly decomposes video and language into multiple structured hierarchies.
arXiv Detail & Related papers (2022-03-24T12:55:23Z) - LexSubCon: Integrating Knowledge from Lexical Resources into Contextual
Embeddings for Lexical Substitution [76.615287796753]
We introduce LexSubCon, an end-to-end lexical substitution framework based on contextual embedding models.
This is achieved by combining contextual information with knowledge from structured lexical resources.
Our experiments show that LexSubCon outperforms previous state-of-the-art methods on LS07 and CoInCo benchmark datasets.
arXiv Detail & Related papers (2021-07-11T21:25:56Z) - Measuring diachronic sense change: new models and Monte Carlo methods
for Bayesian inference [3.1727619150610837]
In a bag-of-words model, the senses of a word with multiple meanings are represented as probability distributions over context words.
We adapt an existing generative sense change model to develop a simpler model for the main effects of sense and time.
We carry out automatic sense-annotation of snippets containing "kosmos" using our model, and measure the time-evolution of its three senses and their prevalence.
arXiv Detail & Related papers (2021-04-14T11:40:21Z) - A Probabilistic Approach in Historical Linguistics Word Order Change in
Infinitival Clauses: from Latin to Old French [0.0]
This thesis investigates word order change in infinitival clauses in the history of Latin and Old French.
I examine a synchronic word order variation in each stage of language change, from which I infer the character, periodization and constraints of diachronic variation.
I present a three-stage probabilistic model of word order change, which also conforms to traditional language change patterns.
arXiv Detail & Related papers (2020-11-16T20:30:31Z) - Grounded Compositional Outputs for Adaptive Language Modeling [59.02706635250856]
A language model's vocabulary$-$typically selected before training and permanently fixed later$-$affects its size.
We propose a fully compositional output embedding layer for language models.
To our knowledge, the result is the first word-level language model with a size that does not depend on the training vocabulary.
arXiv Detail & Related papers (2020-09-24T07:21:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.