Handling Wikidata Qualifiers in Reasoning
- URL: http://arxiv.org/abs/2304.03375v2
- Date: Wed, 21 Jun 2023 13:12:56 GMT
- Title: Handling Wikidata Qualifiers in Reasoning
- Authors: Sahar Aljalbout, Gilles Falquet, Didier Buchs
- Abstract summary: We show how to handle qualifiers in inference rules using Wikidata statements.
We use a many-sorted logical language to formalize the Wikidata model.
We show how to use the MSL and specification to reason on qualifiers.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Wikidata is a knowledge graph increasingly adopted by many communities for
diverse applications. Wikidata statements are annotated with qualifier-value
pairs that are used to depict information, such as the validity context of the
statement, its causality, provenances, etc. Handling the qualifiers in
reasoning is a challenging problem. When defining inference rules (in
particular, rules on ontological properties (x subclass of y, z instance of x,
etc.)), one must consider the qualifiers, as most of them participate in the
semantics of the statements. This poses a complex problem because a) there is a
massive number of qualifiers, and b) the qualifiers of the inferred statement
are often a combination of the qualifiers in the rule condition. In this work,
we propose to address this problem by a) defining a categorization of the
qualifiers b) formalizing the Wikidata model with a many-sorted logical
language; the sorts of this language are the qualifier categories. We couple
this logic with an algebraic specification that provides a means for
effectively handling qualifiers in inference rules. Using Wikidata ontological
properties, we show how to use the MSL and specification to reason on
qualifiers. Finally, we discuss the methodology for practically implementing
the work and present a prototype implementation. The work can be naturally
extended, thanks to the extensibility of the many-sorted algebraic
specification, to cover more qualifiers in the specification, such as uncertain
time, recurring events, geographic locations, and others.
Related papers
- Disjointness Violations in Wikidata [0.0]
We analyze the current modeling of disjointness on Wikidata.
We use SPARQL queries to identify each culprit'' causing a disjointness violation and lay out formulas to identify and fix conflicting information.
arXiv Detail & Related papers (2024-10-17T16:07:51Z) - Learning List-Level Domain-Invariant Representations for Ranking [59.3544317373004]
We propose list-level alignment -- learning domain-invariant representations at the higher level of lists.
The benefits are twofold: it leads to the first domain adaptation generalization bound for ranking, in turn providing theoretical support for the proposed method.
arXiv Detail & Related papers (2022-12-21T04:49:55Z) - CLSE: Corpus of Linguistically Significant Entities [58.29901964387952]
We release a Corpus of Linguistically Significant Entities (CLSE) annotated by experts.
CLSE covers 74 different semantic types to support various applications from airline ticketing to video games.
We create a linguistically representative NLG evaluation benchmark in three languages: French, Marathi, and Russian.
arXiv Detail & Related papers (2022-11-04T12:56:12Z) - Does Wikidata Support Analogical Reasoning? [17.68704739786042]
We investigate whether the knowledge in Wikidata supports analogical reasoning.
We show that Wikidata can be used to create data for analogy classification.
We devise a set of metrics to guide an automatic method for extracting analogies from Wikidata.
arXiv Detail & Related papers (2022-10-02T20:46:52Z) - Multilingual Extraction and Categorization of Lexical Collocations with
Graph-aware Transformers [86.64972552583941]
We put forward a sequence tagging BERT-based model enhanced with a graph-aware transformer architecture, which we evaluate on the task of collocation recognition in context.
Our results suggest that explicitly encoding syntactic dependencies in the model architecture is helpful, and provide insights on differences in collocation typification in English, Spanish and French.
arXiv Detail & Related papers (2022-05-23T16:47:37Z) - Improving Candidate Retrieval with Entity Profile Generation for
Wikidata Entity Linking [76.00737707718795]
We propose a novel candidate retrieval paradigm based on entity profiling.
We use the profile to query the indexed search engine to retrieve candidate entities.
Our approach complements the traditional approach of using a Wikipedia anchor-text dictionary.
arXiv Detail & Related papers (2022-02-27T17:38:53Z) - A Benchmark for Generalizable and Interpretable Temporal Question
Answering over Knowledge Bases [67.33560134350427]
TempQA-WD is a benchmark dataset for temporal reasoning.
It is based on Wikidata, which is the most frequently curated, openly available knowledge base.
arXiv Detail & Related papers (2022-01-15T08:49:09Z) - Survey on English Entity Linking on Wikidata [3.8289963781051415]
Wikidata is a frequently updated, community-driven, and multilingual knowledge graph.
Current Wikidata-specific Entity Linking datasets do not differ in their annotation scheme from schemes for other knowledge graphs like DBpedia.
Almost all approaches employ specific properties like labels and sometimes descriptions but ignore characteristics such as the hyper-relational structure.
arXiv Detail & Related papers (2021-12-03T16:02:42Z) - Mining Knowledge for Natural Language Inference from Wikipedia
Categories [53.26072815839198]
We introduce WikiNLI: a resource for improving model performance on NLI and LE tasks.
It contains 428,899 pairs of phrases constructed from naturally annotated category hierarchies in Wikipedia.
We show that we can improve strong baselines such as BERT and RoBERTa by pretraining them on WikiNLI and transferring the models on downstream tasks.
arXiv Detail & Related papers (2020-10-03T00:45:01Z) - Wikidata on MARS [0.20305676256390934]
Multi-attributed relational structures (MARSs) have been proposed as a formal data model for generalized property graphs.
MARPL is a useful rule-based logic in which to write inference rules over property graphs.
Wikidata can be modelled in an extended MARS that adds the (imprecise) datatypes of Wikidata.
arXiv Detail & Related papers (2020-08-14T22:58:04Z) - Wikidata Constraints on MARS (Extended Technical Report) [0.0]
We show that nearly all Wikidata's existing property constraints can be completely characterized in a natural and economical fashion.
We also give characterizations for several proposed property constraints, and show that a variety of non-property constraints can be handled in the same framework.
arXiv Detail & Related papers (2020-08-10T04:49:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.