Neural-Symbolic Models for Logical Queries on Knowledge Graphs
- URL: http://arxiv.org/abs/2205.10128v1
- Date: Mon, 16 May 2022 18:39:04 GMT
- Title: Neural-Symbolic Models for Logical Queries on Knowledge Graphs
- Authors: Zhaocheng Zhu, Mikhail Galkin, Zuobai Zhang, Jian Tang
- Abstract summary: We propose Graph Neural Network Query Executor (GNN-QE), a neural-symbolic model that enjoys the advantages of both worlds.
GNN-QE decomposes a complex FOL query into relation projections and logical operations over fuzzy sets.
Experiments on 3 datasets show that GNN-QE significantly improves over previous state-of-the-art models in answering FOL queries.
- Score: 17.290758383645567
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Answering complex first-order logic (FOL) queries on knowledge graphs is a
fundamental task for multi-hop reasoning. Traditional symbolic methods traverse
a complete knowledge graph to extract the answers, which provides good
interpretation for each step. Recent neural methods learn geometric embeddings
for complex queries. These methods can generalize to incomplete knowledge
graphs, but their reasoning process is hard to interpret. In this paper, we
propose Graph Neural Network Query Executor (GNN-QE), a neural-symbolic model
that enjoys the advantages of both worlds. GNN-QE decomposes a complex FOL
query into relation projections and logical operations over fuzzy sets, which
provides interpretability for intermediate variables. To reason about the
missing links, GNN-QE adapts a graph neural network from knowledge graph
completion to execute the relation projections, and models the logical
operations with product fuzzy logic. Extensive experiments on 3 datasets show
that GNN-QE significantly improves over previous state-of-the-art models in
answering FOL queries. Meanwhile, GNN-QE can predict the number of answers
without explicit supervision, and provide visualizations for intermediate
variables.
Related papers
- Do graph neural network states contain graph properties? [5.222978725954348]
We present a model explainability pipeline for Graph Neural Networks (GNNs) employing diagnostic classifiers.
This pipeline aims to probe and interpret the learned representations in GNNs across various architectures and datasets.
arXiv Detail & Related papers (2024-11-04T15:26:07Z) - Improving Complex Reasoning over Knowledge Graph with Logic-Aware Curriculum Tuning [89.89857766491475]
We propose a complex reasoning schema over KG upon large language models (LLMs)
We augment the arbitrary first-order logical queries via binary tree decomposition to stimulate the reasoning capability of LLMs.
Experiments across widely used datasets demonstrate that LACT has substantial improvements(brings an average +5.5% MRR score) over advanced methods.
arXiv Detail & Related papers (2024-05-02T18:12:08Z) - Neural Graph Reasoning: Complex Logical Query Answering Meets Graph
Databases [63.96793270418793]
Complex logical query answering (CLQA) is a recently emerged task of graph machine learning.
We introduce the concept of Neural Graph Database (NGDBs)
NGDB consists of a Neural Graph Storage and a Neural Graph Engine.
arXiv Detail & Related papers (2023-03-26T04:03:37Z) - Logical Message Passing Networks with One-hop Inference on Atomic
Formulas [57.47174363091452]
We propose a framework for complex query answering that decomposes the Knowledge Graph embeddings from neural set operators.
On top of the query graph, we propose the Logical Message Passing Neural Network (LMPNN) that connects the local one-hop inferences on atomic formulas to the global logical reasoning.
Our approach yields the new state-of-the-art neural CQA model.
arXiv Detail & Related papers (2023-01-21T02:34:06Z) - Neural Methods for Logical Reasoning Over Knowledge Graphs [14.941769519278745]
We focus on answering multi-hop logical queries on Knowledge Graphs (KGs)
Most previous works have been unable to create models that accept full First-Order Logical (FOL) queries.
We introduce a set of models that use Neural Networks to create one-point vector embeddings to answer the queries.
arXiv Detail & Related papers (2022-09-28T23:10:09Z) - PGX: A Multi-level GNN Explanation Framework Based on Separate Knowledge
Distillation Processes [0.2005299372367689]
We propose a multi-level GNN explanation framework based on an observation that GNN is a multimodal learning process of multiple components in graph data.
The complexity of the original problem is relaxed by breaking into multiple sub-parts represented as a hierarchical structure.
We also aim for personalized explanations as the framework can generate different results based on user preferences.
arXiv Detail & Related papers (2022-08-05T10:14:48Z) - Automatic Relation-aware Graph Network Proliferation [182.30735195376792]
We propose Automatic Relation-aware Graph Network Proliferation (ARGNP) for efficiently searching GNNs.
These operations can extract hierarchical node/relational information and provide anisotropic guidance for message passing on a graph.
Experiments on six datasets for four graph learning tasks demonstrate that GNNs produced by our method are superior to the current state-of-the-art hand-crafted and search-based GNNs.
arXiv Detail & Related papers (2022-05-31T10:38:04Z) - Parameterized Explainer for Graph Neural Network [49.79917262156429]
We propose PGExplainer, a parameterized explainer for Graph Neural Networks (GNNs)
Compared to the existing work, PGExplainer has better generalization ability and can be utilized in an inductive setting easily.
Experiments on both synthetic and real-life datasets show highly competitive performance with up to 24.7% relative improvement in AUC on explaining graph classification.
arXiv Detail & Related papers (2020-11-09T17:15:03Z) - Evaluating Logical Generalization in Graph Neural Networks [59.70452462833374]
We study the task of logical generalization using graph neural networks (GNNs)
Our benchmark suite, GraphLog, requires that learning algorithms perform rule induction in different synthetic logics.
We find that the ability for models to generalize and adapt is strongly determined by the diversity of the logical rules they encounter during training.
arXiv Detail & Related papers (2020-03-14T05:45:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.