Temporal Concept Drift and Alignment: An empirical approach to comparing
Knowledge Organization Systems over time
- URL: http://arxiv.org/abs/2208.07835v1
- Date: Tue, 16 Aug 2022 16:37:17 GMT
- Title: Temporal Concept Drift and Alignment: An empirical approach to comparing
Knowledge Organization Systems over time
- Authors: Sam Grabus (1), Peter Melville Logan (2), Jane Greenberg (1) ((1)
Drexel University, (2) Temple University)
- Abstract summary: This research explores temporal concept drift and temporal alignment in knowledge organization systems (KOS)
A comparative analysis is pursued using the 1910 Library of Congress Subject Headings, 2020 FAST Topical, and automatic indexing.
Results confirm that historical vocabularies can be used to generate anachronistic subject headings representing conceptual drift across time in KOS and historical resources.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This research explores temporal concept drift and temporal alignment in
knowledge organization systems (KOS). A comparative analysis is pursued using
the 1910 Library of Congress Subject Headings, 2020 FAST Topical, and automatic
indexing. The use case involves a sample of 90 nineteenth-century Encyclopedia
Britannica entries. The entries were indexed using two approaches: 1) full-text
indexing; 2) Named Entity Recognition was performed upon the entries with
Stanza, Stanford's NLP toolkit, and entities were automatically indexed with
the Helping Interdisciplinary Vocabulary application (HIVE), using both 1910
LCSH and FAST Topical. The analysis focused on three goals: 1) identifying
results that were exclusive to the 1910 LCSH output; 2) identifying terms in
the exclusive set that have been deprecated from the contemporary LCSH,
demonstrating temporal concept drift; and 3) exploring the historical
significance of these deprecated terms. Results confirm that historical
vocabularies can be used to generate anachronistic subject headings
representing conceptual drift across time in KOS and historical resources. A
methodological contribution is made demonstrating how to study changes in KOS
over time and improve the contextualization of historical humanities resources.
Related papers
- The quasi-semantic competence of LLMs: a case study on the part-whole relation [53.37191762146552]
We investigate knowledge of the emphpart-whole relation, a.k.a. emphmeronymy.
We show that emphquasi-semantic'' models have just a emphquasi-semantic'' competence and still fall short of capturing deep inferential properties.
arXiv Detail & Related papers (2025-04-03T08:41:26Z) - Time Travel: A Comprehensive Benchmark to Evaluate LMMs on Historical and Cultural Artifacts [65.90535970515266]
TimeTravel is a benchmark of 10,250 expert-verified samples spanning 266 distinct cultures across 10 major historical regions.
TimeTravel is designed for AI-driven analysis of manuscripts, artworks, inscriptions, and archaeological discoveries.
We evaluate contemporary AI models on TimeTravel, highlighting their strengths and identifying areas for improvement.
arXiv Detail & Related papers (2025-02-20T18:59:51Z) - Taxonomy-guided Semantic Indexing for Academic Paper Search [51.07749719327668]
TaxoIndex is a semantic index framework for academic paper search.
It organizes key concepts from papers as a semantic index guided by an academic taxonomy.
It can be flexibly employed to enhance existing dense retrievers.
arXiv Detail & Related papers (2024-10-25T00:00:17Z) - From cart to truck: meaning shift through words in English in the last two centuries [0.0]
This onomasiological study uses diachronic word embeddings to explore how different words represented the same concepts over time.
We identify shifts in energy, transport, entertainment, and computing domains, revealing connections between language and societal changes.
arXiv Detail & Related papers (2024-08-29T02:05:39Z) - Data Set Terminology of Deep Learning in Medicine: A Historical Review and Recommendation [0.7897552065199818]
Medicine and deep learning-based artificial intelligence engineering represent two distinct fields each with decades of published history.
With such history comes a set of terminology that has a specific way in which it is applied.
This review aims to give historical context for these terms, accentuate the importance of clarity when these terms are used in medical AI contexts.
arXiv Detail & Related papers (2024-04-30T07:07:45Z) - Language Models As Semantic Indexers [78.83425357657026]
We introduce LMIndexer, a self-supervised framework to learn semantic IDs with a generative language model.
We show the high quality of the learned IDs and demonstrate their effectiveness on three tasks including recommendation, product search, and document retrieval.
arXiv Detail & Related papers (2023-10-11T18:56:15Z) - O-Dang! The Ontology of Dangerous Speech Messages [53.15616413153125]
We present O-Dang!: The Ontology of Dangerous Speech Messages, a systematic and interoperable Knowledge Graph (KG)
O-Dang! is designed to gather and organize Italian datasets into a structured KG, according to the principles shared within the Linguistic Linked Open Data community.
It provides a model for encoding both gold standard and single-annotator labels in the KG.
arXiv Detail & Related papers (2022-07-13T11:50:05Z) - Placing (Historical) Facts on a Timeline: A Classification cum Coref
Resolution Approach [4.809236881780707]
A timeline provides one of the most effective ways to visualize the important historical facts that occurred over a period of time.
We introduce a two staged system for event timeline generation from multiple (historical) text documents.
Our results can be extremely helpful for historians, in advancing research in history and in understanding the socio-political landscape of a country.
arXiv Detail & Related papers (2022-06-28T15:36:44Z) - Supporting Vision-Language Model Inference with Confounder-pruning Knowledge Prompt [71.77504700496004]
Vision-language models are pre-trained by aligning image-text pairs in a common space to deal with open-set visual concepts.
To boost the transferability of the pre-trained models, recent works adopt fixed or learnable prompts.
However, how and what prompts can improve inference performance remains unclear.
arXiv Detail & Related papers (2022-05-23T07:51:15Z) - Overview of the TREC 2020 Fair Ranking Track [64.16623297717642]
This paper provides an overview of the NIST TREC 2020 Fair Ranking track.
The central goal of the Fair Ranking track is to provide fair exposure to different groups of authors.
arXiv Detail & Related papers (2021-08-11T10:22:05Z) - Search from History and Reason for Future: Two-stage Reasoning on
Temporal Knowledge Graphs [56.33651635705633]
We propose CluSTeR to predict future facts in a two-stage manner, Clue Searching and Temporal Reasoning.
CluSTeR learns a beam search policy via reinforcement learning (RL) to induce multiple clues from historical facts.
At the temporal reasoning stage, it adopts a graph convolution network based sequence method to deduce answers from clues.
arXiv Detail & Related papers (2021-06-01T09:01:22Z) - Corpus of Chinese Dynastic Histories: Gender Analysis over Two Millennia [3.2851864672627618]
dynastic histories form a large continuous linguistic space of approximately 2000 years, from the 3rd century BCE to the 18th century CE.
The histories are documented in Classical (Literary) Chinese in a corpus of over 20 million characters, suitable for the computational analysis of historical lexicon and semantic change.
This project introduces a new open-source corpus of twenty-four dynastic histories covered by Creative Commons license.
arXiv Detail & Related papers (2020-05-18T15:14:33Z) - Semantic Relatedness for Keyword Disambiguation: Exploiting Different
Embeddings [0.0]
We propose an approach to keyword disambiguation which grounds on a semantic relatedness between words and senses provided by an external inventory (ontology) that is not known at training time.
Experimental results show that this approach achieves results comparable with the state of the art when applied for Word Sense Disambiguation (WSD) without training for a particular domain.
arXiv Detail & Related papers (2020-02-25T16:44:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.