What is Normal, What is Strange, and What is Missing in a Knowledge
Graph: Unified Characterization via Inductive Summarization
- URL: http://arxiv.org/abs/2003.10412v1
- Date: Mon, 23 Mar 2020 17:38:31 GMT
- Title: What is Normal, What is Strange, and What is Missing in a Knowledge
Graph: Unified Characterization via Inductive Summarization
- Authors: Caleb Belth, Xinyi Zheng, Jilles Vreeken, Danai Koutra
- Abstract summary: We introduce a unified solution to KG characterization by formulating the problem as unsupervised KG summarization.
KGist learns a summary of inductive rules that best compress the KG according to the Minimum Description Length principle.
We show that KGist outperforms task-specific, supervised and unsupervised baselines in error detection and incompleteness identification.
- Score: 34.3446695203147
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Knowledge graphs (KGs) store highly heterogeneous information about the world
in the structure of a graph, and are useful for tasks such as question
answering and reasoning. However, they often contain errors and are missing
information. Vibrant research in KG refinement has worked to resolve these
issues, tailoring techniques to either detect specific types of errors or
complete a KG.
In this work, we introduce a unified solution to KG characterization by
formulating the problem as unsupervised KG summarization with a set of
inductive, soft rules, which describe what is normal in a KG, and thus can be
used to identify what is abnormal, whether it be strange or missing. Unlike
first-order logic rules, our rules are labeled, rooted graphs, i.e., patterns
that describe the expected neighborhood around a (seen or unseen) node, based
on its type, and information in the KG. Stepping away from the traditional
support/confidence-based rule mining techniques, we propose KGist, Knowledge
Graph Inductive SummarizaTion, which learns a summary of inductive rules that
best compress the KG according to the Minimum Description Length principle---a
formulation that we are the first to use in the context of KG rule mining. We
apply our rules to three large KGs (NELL, DBpedia, and Yago), and tasks such as
compression, various types of error detection, and identification of incomplete
information. We show that KGist outperforms task-specific, supervised and
unsupervised baselines in error detection and incompleteness identification,
(identifying the location of up to 93% of missing entities---over 10% more than
baselines), while also being efficient for large knowledge graphs.
Related papers
- A Prompt-Based Knowledge Graph Foundation Model for Universal In-Context Reasoning [17.676185326247946]
We propose a prompt-based KG foundation model via in-context learning, namely KG-ICL, to achieve a universal reasoning ability.
To encode prompt graphs with the generalization ability to unseen entities and relations in queries, we first propose a unified tokenizer.
Then, we propose two message passing neural networks to perform prompt encoding and KG reasoning, respectively.
arXiv Detail & Related papers (2024-10-16T06:47:18Z) - Learning Rules from KGs Guided by Language Models [48.858741745144044]
Rule learning methods can be applied to predict potentially missing facts.
Ranking of rules is especially challenging over highly incomplete or biased KGs.
With the recent rise of Language Models (LMs) several works have claimed that LMs can be used as alternative means for KG completion.
arXiv Detail & Related papers (2024-09-12T09:27:36Z) - Generate-on-Graph: Treat LLM as both Agent and KG in Incomplete Knowledge Graph Question Answering [87.67177556994525]
We propose a training-free method called Generate-on-Graph (GoG) to generate new factual triples while exploring Knowledge Graphs (KGs)
GoG performs reasoning through a Thinking-Searching-Generating framework, which treats LLM as both Agent and KG in IKGQA.
arXiv Detail & Related papers (2024-04-23T04:47:22Z) - ChatRule: Mining Logical Rules with Large Language Models for Knowledge
Graph Reasoning [107.61997887260056]
We propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs.
Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs.
To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs.
arXiv Detail & Related papers (2023-09-04T11:38:02Z) - Contrastive Knowledge Graph Error Detection [11.637359888052014]
We propose a novel framework - ContrAstive knowledge Graph Error Detection (CAGED)
CAGED introduces contrastive learning into KG learning and provides a novel way of modeling KG.
It outperforms state-of-the-art methods in KG error detection.
arXiv Detail & Related papers (2022-11-18T05:01:19Z) - Explainable Sparse Knowledge Graph Completion via High-order Graph
Reasoning Network [111.67744771462873]
This paper proposes a novel explainable model for sparse Knowledge Graphs (KGs)
It combines high-order reasoning into a graph convolutional network, namely HoGRN.
It can not only improve the generalization ability to mitigate the information insufficiency issue but also provide interpretability.
arXiv Detail & Related papers (2022-07-14T10:16:56Z) - MEKER: Memory Efficient Knowledge Embedding Representation for Link
Prediction and Question Answering [65.62309538202771]
Knowledge Graphs (KGs) are symbolically structured storages of facts.
KG embedding contains concise data used in NLP tasks requiring implicit information about the real world.
We propose a memory-efficient KG embedding model, which yields SOTA-comparable performance on link prediction tasks and KG-based Question Answering.
arXiv Detail & Related papers (2022-04-22T10:47:03Z) - DegreEmbed: incorporating entity embedding into logic rule learning for
knowledge graph reasoning [7.066269573204757]
Link prediction for knowledge graphs is the task aiming to complete missing facts by reasoning based on the existing knowledge.
We propose DegreEmbed, a model that combines embedding-based learning and logic rule mining for inferring on KGs.
arXiv Detail & Related papers (2021-12-18T13:38:48Z) - Entity Type Prediction in Knowledge Graphs using Embeddings [2.7528170226206443]
Open Knowledge Graphs (such as DBpedia, Wikidata, YAGO) have been recognized as the backbone of diverse applications in the field of data mining and information retrieval.
Most of these KGs are mostly created either via an automated information extraction from snapshots or information accumulation provided by the users or using Wikipedias.
It has been observed that the type information of these KGs is often noisy, incomplete, and incorrect.
A multi-label classification approach is proposed in this work for entity typing using KG embeddings.
arXiv Detail & Related papers (2020-04-28T17:57:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.