Semantic coordinates analysis reveals language changes in the AI field
- URL: http://arxiv.org/abs/2011.00543v1
- Date: Sun, 1 Nov 2020 15:59:24 GMT
- Title: Semantic coordinates analysis reveals language changes in the AI field
- Authors: Zining Zhu, Yang Xu, Frank Rudzicz
- Abstract summary: We propose a method based on semantic shifts that reveals changes in language within publications of a field.
We use GloVe-style probability ratios to quantify the shifting directions and extents from multiple viewpoints.
We show that semantic coordinates analysis can detect shifts echoing changes of research interests.
- Score: 19.878987032985634
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semantic shifts can reflect changes in beliefs across hundreds of years, but
it is less clear whether trends in fast-changing communities across a short
time can be detected. We propose semantic coordinates analysis, a method based
on semantic shifts, that reveals changes in language within publications of a
field (we use AI as example) across a short time span. We use GloVe-style
probability ratios to quantify the shifting directions and extents from
multiple viewpoints. We show that semantic coordinates analysis can detect
shifts echoing changes of research interests (e.g., "deep" shifted further from
"rigorous" to "neural"), and developments of research activities (e,g.,
"collaboration" contains less "competition" than "collaboration"), based on
publications spanning as short as 10 years.
Related papers
- Survey in Characterization of Semantic Change [0.1474723404975345]
Understanding the meaning of words is vital for interpreting texts from different cultures.
Semantic changes can potentially impact the quality of the outcomes of computational linguistics algorithms.
arXiv Detail & Related papers (2024-02-29T12:13:50Z) - Syntactic Language Change in English and German: Metrics, Parsers, and Convergences [56.47832275431858]
The current paper looks at diachronic trends in syntactic language change in both English and German, using corpora of parliamentary debates from the last c. 160 years.
We base our observations on five dependencys, including the widely used Stanford Core as well as 4 newer alternatives.
We show that changes in syntactic measures seem to be more frequent at the tails of sentence length distributions.
arXiv Detail & Related papers (2024-02-18T11:46:16Z) - Towards Unsupervised Recognition of Token-level Semantic Differences in
Related Documents [61.63208012250885]
We formulate recognizing semantic differences as a token-level regression task.
We study three unsupervised approaches that rely on a masked language model.
Our results show that an approach based on word alignment and sentence-level contrastive learning has a robust correlation to gold labels.
arXiv Detail & Related papers (2023-05-22T17:58:04Z) - The Problem of Semantic Shift in Longitudinal Monitoring of Social
Media: A Case Study on Mental Health During the COVID-19 Pandemic [15.002282686061905]
Social media allows researchers to track societal and cultural changes over time based on language analysis tools.
Many of these tools rely on statistical algorithms which need to be tuned to specific types of language.
Recent studies have shown the absence of appropriate tuning, specifically in the presence of semantic shift, can hinder robustness of the underlying methods.
arXiv Detail & Related papers (2022-06-22T15:09:28Z) - Compositional Temporal Grounding with Structured Variational Cross-Graph
Correspondence Learning [92.07643510310766]
Temporal grounding in videos aims to localize one target video segment that semantically corresponds to a given query sentence.
We introduce a new Compositional Temporal Grounding task and construct two new dataset splits.
We empirically find that they fail to generalize to queries with novel combinations of seen words.
We propose a variational cross-graph reasoning framework that explicitly decomposes video and language into multiple structured hierarchies.
arXiv Detail & Related papers (2022-03-24T12:55:23Z) - SCoT: Sense Clustering over Time: a tool for the analysis of lexical
change [79.80787569986283]
We present Sense Clustering over Time (SCoT), a novel network-based tool for analysing lexical change.
SCoT represents the meanings of a word as clusters of similar words.
It has been successfully used in a European study on the changing meaning of crisis'
arXiv Detail & Related papers (2022-03-18T12:04:09Z) - Contextualized Semantic Distance between Highly Overlapped Texts [85.1541170468617]
Overlapping frequently occurs in paired texts in natural language processing tasks like text editing and semantic similarity evaluation.
This paper aims to address the issue with a mask-and-predict strategy.
We take the words in the longest common sequence as neighboring words and use masked language modeling (MLM) to predict the distributions on their positions.
Experiments on Semantic Textual Similarity show NDD to be more sensitive to various semantic differences, especially on highly overlapped paired texts.
arXiv Detail & Related papers (2021-10-04T03:59:15Z) - Lexical Semantic Change Discovery [22.934650688233734]
We propose a shift from change detection to change discovery, i.e., discovering novel word senses over time from the full corpus vocabulary.
By heavily fine-tuning a type-based and a token-based approach on recently published German data, we demonstrate that both models can successfully be applied to discover new words undergoing meaning change.
arXiv Detail & Related papers (2021-06-06T13:02:38Z) - Fake it Till You Make it: Self-Supervised Semantic Shifts for
Monolingual Word Embedding Tasks [58.87961226278285]
We propose a self-supervised approach to model lexical semantic change.
We show that our method can be used for the detection of semantic change with any alignment method.
We illustrate the utility of our techniques using experimental results on three different datasets.
arXiv Detail & Related papers (2021-01-30T18:59:43Z) - Capturing Evolution in Word Usage: Just Add More Clusters? [9.209873675320834]
We focus on a new set of methods relying on contextualised embeddings, a type of semantic modelling that revolutionised the NLP field recently.
We leverage the ability of the transformer-based BERT model to generate contextualised embeddings capable of detecting semantic change of words across time.
arXiv Detail & Related papers (2020-01-18T09:04:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.