Measuring diachronic sense change: new models and Monte Carlo methods
for Bayesian inference
- URL: http://arxiv.org/abs/2105.00819v1
- Date: Wed, 14 Apr 2021 11:40:21 GMT
- Title: Measuring diachronic sense change: new models and Monte Carlo methods
for Bayesian inference
- Authors: Schyan Zafar and Geoff Nicholls
- Abstract summary: In a bag-of-words model, the senses of a word with multiple meanings are represented as probability distributions over context words.
We adapt an existing generative sense change model to develop a simpler model for the main effects of sense and time.
We carry out automatic sense-annotation of snippets containing "kosmos" using our model, and measure the time-evolution of its three senses and their prevalence.
- Score: 3.1727619150610837
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In a bag-of-words model, the senses of a word with multiple meanings, e.g.
"bank" (used either in a river-bank or an institution sense), are represented
as probability distributions over context words, and sense prevalence is
represented as a probability distribution over senses. Both of these may change
with time. Modelling and measuring this kind of sense change is challenging due
to the typically high-dimensional parameter space and sparse datasets. A
recently published corpus of ancient Greek texts contains expert-annotated
sense labels for selected target words. Automatic sense-annotation for the word
"kosmos" (meaning decoration, order or world) has been used as a test case in
recent work with related generative models and Monte Carlo methods. We adapt an
existing generative sense change model to develop a simpler model for the main
effects of sense and time, and give MCMC methods for Bayesian inference on all
these models that are more efficient than existing methods. We carry out
automatic sense-annotation of snippets containing "kosmos" using our model, and
measure the time-evolution of its three senses and their prevalence. As far as
we are aware, ours is the first analysis of this data, within the class of
generative models we consider, that quantifies uncertainty and returns credible
sets for evolving sense prevalence in good agreement with those given by expert
annotation.
Related papers
- Human-in-the-loop: Towards Label Embeddings for Measuring Classification Difficulty [14.452983136429967]
In supervised learning, uncertainty can already occur in the first stage of the training process, the annotation phase.
The main idea of this work is to drop the assumption of a ground truth label and instead embed the annotations into a multidimensional space.
The methods developed in this paper readily extend to various situations where multiple annotators independently label instances.
arXiv Detail & Related papers (2023-11-15T11:23:15Z) - An Embedded Diachronic Sense Change Model with a Case Study from Ancient Greek [0.4143603294943439]
This paper introduces EDiSC, an Embedded model, which combines word embeddings with languages to provide superior model performance.
It is shown empirically that EDiSC offers improved accuracy, ground-truth recovery and uncertainty quantification.
arXiv Detail & Related papers (2023-11-01T14:20:18Z) - On the Efficacy of Sampling Adapters [82.5941326570812]
We propose a unified framework for understanding sampling adapters.
We argue that the shift they enforce can be viewed as a trade-off between precision and recall.
We find that several precision-emphasizing measures indeed indicate that sampling adapters can lead to probability distributions more aligned with the true distribution.
arXiv Detail & Related papers (2023-07-07T17:59:12Z) - On the Strong Correlation Between Model Invariance and Generalization [54.812786542023325]
Generalization captures a model's ability to classify unseen data.
Invariance measures consistency of model predictions on transformations of the data.
From a dataset-centric view, we find a certain model's accuracy and invariance linearly correlated on different test sets.
arXiv Detail & Related papers (2022-07-14T17:08:25Z) - A moment-matching metric for latent variable generative models [0.0]
In scope of Goodhart's law, when a metric becomes a target it ceases to be a good metric.
We propose a new metric for model comparison or regularization that relies on moments.
It is common to draw samples from the fitted distribution when evaluating latent variable models.
arXiv Detail & Related papers (2021-10-04T17:51:08Z) - A comprehensive comparative evaluation and analysis of Distributional
Semantic Models [61.41800660636555]
We perform a comprehensive evaluation of type distributional vectors, either produced by static DSMs or obtained by averaging the contextualized vectors generated by BERT.
The results show that the alleged superiority of predict based models is more apparent than real, and surely not ubiquitous.
We borrow from cognitive neuroscience the methodology of Representational Similarity Analysis (RSA) to inspect the semantic spaces generated by distributional models.
arXiv Detail & Related papers (2021-05-20T15:18:06Z) - Lexical semantic change for Ancient Greek and Latin [61.69697586178796]
Associating a word's correct meaning in its historical context is a central challenge in diachronic research.
We build on a recent computational approach to semantic change based on a dynamic Bayesian mixture model.
We provide a systematic comparison of dynamic Bayesian mixture models for semantic change with state-of-the-art embedding-based models.
arXiv Detail & Related papers (2021-01-22T12:04:08Z) - MASKER: Masked Keyword Regularization for Reliable Text Classification [73.90326322794803]
We propose a fine-tuning method, coined masked keyword regularization (MASKER), that facilitates context-based prediction.
MASKER regularizes the model to reconstruct the keywords from the rest of the words and make low-confidence predictions without enough context.
We demonstrate that MASKER improves OOD detection and cross-domain generalization without degrading classification accuracy.
arXiv Detail & Related papers (2020-12-17T04:54:16Z) - SChME at SemEval-2020 Task 1: A Model Ensemble for Detecting Lexical
Semantic Change [58.87961226278285]
This paper describes SChME, a method used in SemEval-2020 Task 1 on unsupervised detection of lexical semantic change.
SChME usesa model ensemble combining signals of distributional models (word embeddings) and wordfrequency models where each model casts a vote indicating the probability that a word sufferedsemantic change according to that feature.
arXiv Detail & Related papers (2020-12-02T23:56:34Z) - Modelling Lexical Ambiguity with Density Matrices [3.7692411550925664]
We present three new neural models for learning density matrices from a corpus.
Test their ability to discriminate between word senses on a range of compositional datasets.
arXiv Detail & Related papers (2020-10-12T13:08:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.