Knowledge Graphs are not Created Equal: Exploring the Properties and
Structure of Real KGs
- URL: http://arxiv.org/abs/2311.06414v1
- Date: Fri, 10 Nov 2023 22:18:09 GMT
- Title: Knowledge Graphs are not Created Equal: Exploring the Properties and
Structure of Real KGs
- Authors: Nedelina Teneva and Estevam Hruschka
- Abstract summary: We study 29 real knowledge graph datasets from diverse domains to analyze their properties and structural patterns.
We believe that the rich structural information contained in KGs can benefit the development of better KG models across fields.
- Score: 2.28438857884398
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Despite the recent popularity of knowledge graph (KG) related tasks and
benchmarks such as KG embeddings, link prediction, entity alignment and
evaluation of the reasoning abilities of pretrained language models as KGs, the
structure and properties of real KGs are not well studied. In this paper, we
perform a large scale comparative study of 29 real KG datasets from diverse
domains such as the natural sciences, medicine, and NLP to analyze their
properties and structural patterns. Based on our findings, we make several
recommendations regarding KG-based model development and evaluation. We believe
that the rich structural information contained in KGs can benefit the
development of better KG models across fields and we hope this study will
contribute to breaking the existing data silos between different areas of
research (e.g., ML, NLP, AI for sciences).
Related papers
- Distill-SynthKG: Distilling Knowledge Graph Synthesis Workflow for Improved Coverage and Efficiency [59.6772484292295]
Knowledge graphs (KGs) generated by large language models (LLMs) are increasingly valuable for Retrieval-Augmented Generation (RAG) applications.
Existing KG extraction methods rely on prompt-based approaches, which are inefficient for processing large-scale corpora.
We propose SynthKG, a multi-step, document-level synthesis KG workflow based on LLMs.
We also design a novel graph-based retrieval framework for RAG.
arXiv Detail & Related papers (2024-10-22T00:47:54Z) - KG-FIT: Knowledge Graph Fine-Tuning Upon Open-World Knowledge [63.19837262782962]
Knowledge Graph Embedding (KGE) techniques are crucial in learning compact representations of entities and relations within a knowledge graph.
This study introduces KG-FIT, which builds a semantically coherent hierarchical structure of entity clusters.
Experiments on the benchmark datasets FB15K-237, YAGO3-10, and PrimeKG demonstrate the superiority of KG-FIT over state-of-the-art pre-trained language model-based methods.
arXiv Detail & Related papers (2024-05-26T03:04:26Z) - KG-GPT: A General Framework for Reasoning on Knowledge Graphs Using
Large Language Models [18.20425100517317]
We propose KG-GPT, a framework leveraging large language models for tasks employing knowledge graphs.
KG-GPT comprises three steps: Sentence, Graph Retrieval, and Inference, each aimed at partitioning sentences, retrieving relevant graph components, and deriving logical conclusions.
We evaluate KG-GPT using KG-based fact verification and KGQA benchmarks, with the model showing competitive and robust performance, even outperforming several fully-supervised models.
arXiv Detail & Related papers (2023-10-17T12:51:35Z) - LLMs for Knowledge Graph Construction and Reasoning: Recent Capabilities and Future Opportunities [66.36633042421387]
Large Language Models (LLMs) for Knowledge Graph (KG) construction and reasoning evaluated.
We propose AutoKG, a multi-agent-based approach employing LLMs and external sources for KG construction and reasoning.
arXiv Detail & Related papers (2023-05-22T15:56:44Z) - Construction of Knowledge Graphs: State and Challenges [2.245333517888782]
We discuss the main graph models for knowledge graphs (KGs) and introduce the major requirement for future KG construction pipelines.
Next, we provide an overview of the necessary steps to build high-quality KGs, including cross-cutting topics such as metadata management.
We evaluate the state of the art of KG construction w.r.t the introduced requirements for specific popular KGs as well as some recent tools and strategies for KG construction.
arXiv Detail & Related papers (2023-02-22T17:26:03Z) - Reasoning over Multi-view Knowledge Graphs [59.99051368907095]
ROMA is a novel framework for answering logical queries over multi-view KGs.
It scales up to KGs of large sizes (e.g., millions of facts) and fine-granular views.
It generalizes to query structures and KG views that are unobserved during training.
arXiv Detail & Related papers (2022-09-27T21:32:20Z) - BertNet: Harvesting Knowledge Graphs with Arbitrary Relations from
Pretrained Language Models [65.51390418485207]
We propose a new approach of harvesting massive KGs of arbitrary relations from pretrained LMs.
With minimal input of a relation definition, the approach efficiently searches in the vast entity pair space to extract diverse accurate knowledge.
We deploy the approach to harvest KGs of over 400 new relations from different LMs.
arXiv Detail & Related papers (2022-06-28T19:46:29Z) - Meta-Learning Based Knowledge Extrapolation for Knowledge Graphs in the
Federated Setting [43.85991094675398]
We study the knowledge extrapolation problem to embed new components (i.e., entities and relations) that come with emerging knowledge graphs (KGs) in the federated setting.
In this problem, a model trained on an existing KG needs to embed an emerging KG with unseen entities and relations.
We introduce the meta-learning setting, where a set of tasks are sampled on the existing KG to mimic the link prediction task on the emerging KG.
Based on sampled tasks, we meta-train a graph neural network framework that can construct features for unseen components based on structural information and output embeddings for them.
arXiv Detail & Related papers (2022-05-10T06:27:32Z) - Scientific Language Models for Biomedical Knowledge Base Completion: An
Empirical Study [62.376800537374024]
We study scientific LMs for KG completion, exploring whether we can tap into their latent knowledge to enhance biomedical link prediction.
We integrate the LM-based models with KG embedding models, using a router method that learns to assign each input example to either type of model and provides a substantial boost in performance.
arXiv Detail & Related papers (2021-06-17T17:55:33Z) - Learning to Deceive Knowledge Graph Augmented Models via Targeted
Perturbation [42.407209719347286]
Knowledge graphs (KGs) have helped neural models improve performance on various knowledge-intensive tasks.
We show that, through a reinforcement learning policy, one can produce deceptively perturbed KGs.
Our findings raise doubts about KG-augmented models' ability to reason about KG information and give sensible explanations.
arXiv Detail & Related papers (2020-10-24T11:04:45Z) - Relational Learning Analysis of Social Politics using Knowledge Graph
Embedding [11.978556412301975]
This paper presents a novel credibility domain-based KG Embedding framework.
It involves capturing a fusion of data obtained from heterogeneous resources into a formal KG representation depicted by a domain.
The framework also embodies a credibility module to ensure data quality and trustworthiness.
arXiv Detail & Related papers (2020-06-02T14:10:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.