AdaProp: Learning Adaptive Propagation for Graph Neural Network based
Knowledge Graph Reasoning
- URL: http://arxiv.org/abs/2205.15319v2
- Date: Wed, 14 Jun 2023 03:33:04 GMT
- Title: AdaProp: Learning Adaptive Propagation for Graph Neural Network based
Knowledge Graph Reasoning
- Authors: Yongqi Zhang, Zhanke Zhou, Quanming Yao, Xiaowen Chu, Bo Han
- Abstract summary: An important design component of GNN-based reasoning methods is called the propagation path.
We learn an adaptive propagation path in order to filter out irrelevant entities while preserving promising targets.
Our method is powerful, efficient, and semantic-aware.
- Score: 43.06729402877713
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Due to the popularity of Graph Neural Networks (GNNs), various GNN-based
methods have been designed to reason on knowledge graphs (KGs). An important
design component of GNN-based KG reasoning methods is called the propagation
path, which contains a set of involved entities in each propagation step.
Existing methods use hand-designed propagation paths, ignoring the correlation
between the entities and the query relation. In addition, the number of
involved entities will explosively grow at larger propagation steps. In this
work, we are motivated to learn an adaptive propagation path in order to filter
out irrelevant entities while preserving promising targets. First, we design an
incremental sampling mechanism where the nearby targets and layer-wise
connections can be preserved with linear complexity. Second, we design a
learning-based sampling distribution to identify the semantically related
entities. Extensive experiments show that our method is powerful, efficient,
and semantic-aware. The code is available at
https://github.com/LARS-research/AdaProp.
Related papers
- Leveraging Multi-facet Paths for Heterogeneous Graph Representation Learning [2.603958690885184]
We introduce MF2Vec, a model that uses multi-faceted (fine-grained) paths instead of predefined meta-paths.
This method learns diverse aspects of nodes and their relationships, constructs a homogeneous network, and creates node embeddings for classification, link prediction, and clustering.
arXiv Detail & Related papers (2024-07-30T08:45:32Z) - Automatic Relation-aware Graph Network Proliferation [182.30735195376792]
We propose Automatic Relation-aware Graph Network Proliferation (ARGNP) for efficiently searching GNNs.
These operations can extract hierarchical node/relational information and provide anisotropic guidance for message passing on a graph.
Experiments on six datasets for four graph learning tasks demonstrate that GNNs produced by our method are superior to the current state-of-the-art hand-crafted and search-based GNNs.
arXiv Detail & Related papers (2022-05-31T10:38:04Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Graph Neural Networks with Feature and Structure Aware Random Walk [7.143879014059894]
We show that in typical heterphilous graphs, the edges may be directed, and whether to treat the edges as is or simply make them undirected greatly affects the performance of the GNN models.
We develop a model that adaptively learns the directionality of the graph, and exploits the underlying long-distance correlations between nodes.
arXiv Detail & Related papers (2021-11-19T08:54:21Z) - AdaGNN: A multi-modal latent representation meta-learner for GNNs based
on AdaBoosting [0.38073142980733]
Graph Neural Networks (GNNs) focus on extracting intrinsic network features.
We propose boosting-based meta learner for GNNs.
AdaGNN performs exceptionally well for applications with rich and diverse node neighborhood information.
arXiv Detail & Related papers (2021-08-14T03:07:26Z) - ROD: Reception-aware Online Distillation for Sparse Graphs [23.55530524584572]
We propose ROD, a novel reception-aware online knowledge distillation approach for sparse graph learning.
We design three supervision signals for ROD: multi-scale reception-aware graph knowledge, task-based supervision, and rich distilled knowledge.
Our approach has been extensively evaluated on 9 datasets and a variety of graph-based tasks.
arXiv Detail & Related papers (2021-07-25T11:55:47Z) - Higher-Order Attribute-Enhancing Heterogeneous Graph Neural Networks [67.25782890241496]
We propose a higher-order Attribute-Enhancing Graph Neural Network (HAEGNN) for heterogeneous network representation learning.
HAEGNN simultaneously incorporates meta-paths and meta-graphs for rich, heterogeneous semantics.
It shows superior performance against the state-of-the-art methods in node classification, node clustering, and visualization.
arXiv Detail & Related papers (2021-04-16T04:56:38Z) - Learning Intents behind Interactions with Knowledge Graph for
Recommendation [93.08709357435991]
Knowledge graph (KG) plays an increasingly important role in recommender systems.
Existing GNN-based models fail to identify user-item relation at a fine-grained level of intents.
We propose a new model, Knowledge Graph-based Intent Network (KGIN)
arXiv Detail & Related papers (2021-02-14T03:21:36Z) - Learning the Implicit Semantic Representation on Graph-Structured Data [57.670106959061634]
Existing representation learning methods in graph convolutional networks are mainly designed by describing the neighborhood of each node as a perceptual whole.
We propose a Semantic Graph Convolutional Networks (SGCN) that explores the implicit semantics by learning latent semantic-paths in graphs.
arXiv Detail & Related papers (2021-01-16T16:18:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.