Wikidata on MARS
- URL: http://arxiv.org/abs/2008.06599v1
- Date: Fri, 14 Aug 2020 22:58:04 GMT
- Title: Wikidata on MARS
- Authors: Peter F. Patel-Schneider and David Martin
- Abstract summary: Multi-attributed relational structures (MARSs) have been proposed as a formal data model for generalized property graphs.
MARPL is a useful rule-based logic in which to write inference rules over property graphs.
Wikidata can be modelled in an extended MARS that adds the (imprecise) datatypes of Wikidata.
- Score: 0.20305676256390934
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-attributed relational structures (MARSs) have been proposed as a formal
data model for generalized property graphs, along with multi-attributed
rule-based predicate logic (MARPL) as a useful rule-based logic in which to
write inference rules over property graphs. Wikidata can be modelled in an
extended MARS that adds the (imprecise) datatypes of Wikidata. The rules of
inference for the Wikidata ontology can be modelled as a MARPL ontology, with
extensions to handle the Wikidata datatypes and functions over these datatypes.
Because many Wikidata qualifiers should participate in most inference rules in
Wikidata a method of implicitly handling qualifier values on a per-qualifier
basis is needed to make this modelling useful. The meaning of Wikidata is then
the extended MARS that is the closure of running these rules on the Wikidata
data model. Wikidata constraints can be modelled as multi-attributed predicate
logic (MAPL) formulae, again extended with datatypes, that are evaluated over
this extended MARS. The result models Wikidata in a way that fixes several of
its major problems.
Related papers
- Less is More: Making Smaller Language Models Competent Subgraph Retrievers for Multi-hop KGQA [51.3033125256716]
We model the subgraph retrieval task as a conditional generation task handled by small language models.
Our base generative subgraph retrieval model, consisting of only 220M parameters, competitive retrieval performance compared to state-of-the-art models.
Our largest 3B model, when plugged with an LLM reader, sets new SOTA end-to-end performance on both the WebQSP and CWQ benchmarks.
arXiv Detail & Related papers (2024-10-08T15:22:36Z) - Towards a Brazilian History Knowledge Graph [50.26735825937335]
We construct a knowledge graph for Brazilian history based on the Brazilian Dictionary of Historical Biographies (DHBB) and Wikipedia/Wikidata.
We show that many terms/entities described in the DHBB do not have corresponding concepts (or Q items) in Wikidata.
arXiv Detail & Related papers (2024-03-28T22:05:32Z) - KIF: A Wikidata-Based Framework for Integrating Heterogeneous Knowledge Sources [0.45141207783683707]
We present a Wikidata-based framework, called KIF, for virtually integrating heterogeneous knowledge sources.
KIF is written in Python and is released as open-source.
arXiv Detail & Related papers (2024-03-15T13:46:36Z) - Mapping Process for the Task: Wikidata Statements to Text as Wikipedia
Sentences [68.8204255655161]
We propose our mapping process for the task of converting Wikidata statements to natural language text (WS2T) for Wikipedia projects at the sentence level.
The main step is to organize statements, represented as a group of quadruples and triples, and then to map them to corresponding sentences in English Wikipedia.
We evaluate the output corpus in various aspects: sentence structure analysis, noise filtering, and relationships between sentence components based on word embedding models.
arXiv Detail & Related papers (2022-10-23T08:34:33Z) - Does Wikidata Support Analogical Reasoning? [17.68704739786042]
We investigate whether the knowledge in Wikidata supports analogical reasoning.
We show that Wikidata can be used to create data for analogy classification.
We devise a set of metrics to guide an automatic method for extracting analogies from Wikidata.
arXiv Detail & Related papers (2022-10-02T20:46:52Z) - Enriching Wikidata with Linked Open Data [4.311189028205597]
Current linked open data (LOD) tools are not suitable to enrich large graphs like Wikidata.
We present a novel workflow that includes gap detection, source selection, schema alignment, and semantic validation.
Our experiments show that our workflow can enrich Wikidata with millions of novel statements from external LOD sources with a high quality.
arXiv Detail & Related papers (2022-07-01T01:50:24Z) - Improving Candidate Retrieval with Entity Profile Generation for
Wikidata Entity Linking [76.00737707718795]
We propose a novel candidate retrieval paradigm based on entity profiling.
We use the profile to query the indexed search engine to retrieve candidate entities.
Our approach complements the traditional approach of using a Wikipedia anchor-text dictionary.
arXiv Detail & Related papers (2022-02-27T17:38:53Z) - Survey on English Entity Linking on Wikidata [3.8289963781051415]
Wikidata is a frequently updated, community-driven, and multilingual knowledge graph.
Current Wikidata-specific Entity Linking datasets do not differ in their annotation scheme from schemes for other knowledge graphs like DBpedia.
Almost all approaches employ specific properties like labels and sometimes descriptions but ignore characteristics such as the hyper-relational structure.
arXiv Detail & Related papers (2021-12-03T16:02:42Z) - Assessing the quality of sources in Wikidata across languages: a hybrid
approach [64.05097584373979]
We run a series of microtasks experiments to evaluate a large corpus of references, sampled from Wikidata triples with labels in several languages.
We use a consolidated, curated version of the crowdsourced assessments to train several machine learning models to scale up the analysis to the whole of Wikidata.
The findings help us ascertain the quality of references in Wikidata, and identify common challenges in defining and capturing the quality of user-generated multilingual structured data on the web.
arXiv Detail & Related papers (2021-09-20T10:06:46Z) - Creating and Querying Personalized Versions of Wikidata on a Laptop [0.7449724123186383]
This paper introduces KGTK Kypher, a query language and processor that allows users to create personalized variants of Wikidata on a laptop.
We present several use cases that illustrate the types of analyses that Kypher enables users to run on the full Wikidata KG on a laptop.
arXiv Detail & Related papers (2021-08-06T00:00:33Z) - Partially-Aligned Data-to-Text Generation with Distant Supervision [69.15410325679635]
We propose a new generation task called Partially-Aligned Data-to-Text Generation (PADTG)
It is more practical since it utilizes automatically annotated data for training and thus considerably expands the application domains.
Our framework outperforms all baseline models as well as verify the feasibility of utilizing partially-aligned data.
arXiv Detail & Related papers (2020-10-03T03:18:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.