DsMtGCN: A Direction-sensitive Multi-task framework for Knowledge Graph
Completion
- URL: http://arxiv.org/abs/2306.10290v1
- Date: Sat, 17 Jun 2023 08:21:47 GMT
- Title: DsMtGCN: A Direction-sensitive Multi-task framework for Knowledge Graph
Completion
- Authors: Jining Wang, Chuan Chen, Zibin Zheng, Yuren Zhou
- Abstract summary: We propose a Direction-sensitive Multi-task GCN (DsMtGCN) to make full use of the direction information.
In this paper, we propose a Direction-sensitive Multi-task GCN (DsMtGCN) to make full use of the direction information.
- Score: 31.102735901710563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To solve the inherent incompleteness of knowledge graphs (KGs), numbers of
knowledge graph completion (KGC) models have been proposed to predict missing
links from known triples. Among those, several works have achieved more
advanced results via exploiting the structure information on KGs with Graph
Convolutional Networks (GCN). However, we observe that entity embeddings
aggregated from neighbors in different directions are just simply averaged to
complete single-tasks by existing GCN based models, ignoring the specific
requirements of forward and backward sub-tasks. In this paper, we propose a
Direction-sensitive Multi-task GCN (DsMtGCN) to make full use of the direction
information, the multi-head self-attention is applied to specifically combine
embeddings in different directions based on various entities and sub-tasks, the
geometric constraints are imposed to adjust the distribution of embeddings, and
the traditional binary cross-entropy loss is modified to reflect the triple
uncertainty. Moreover, the competitive experiments results on several benchmark
datasets verify the effectiveness of our model.
Related papers
- Beyond Entity Alignment: Towards Complete Knowledge Graph Alignment via Entity-Relation Synergy [14.459419325027612]
Knowledge Graph alignment aims to integrate knowledge from multiple sources to address the limitations of individual Knowledge Graphs.
Existing models primarily emphasize the linkage of cross-graph entities but overlook aligning relations across KGs.
We propose a novel Expectation-Maximization-based model, EREM, which iteratively optimize both sub-tasks.
arXiv Detail & Related papers (2024-07-25T03:40:09Z) - ARC: A Generalist Graph Anomaly Detector with In-Context Learning [62.202323209244]
ARC is a generalist GAD approach that enables a one-for-all'' GAD model to detect anomalies across various graph datasets on-the-fly.
equipped with in-context learning, ARC can directly extract dataset-specific patterns from the target dataset.
Extensive experiments on multiple benchmark datasets from various domains demonstrate the superior anomaly detection performance, efficiency, and generalizability of ARC.
arXiv Detail & Related papers (2024-05-27T02:42:33Z) - ULTRA-DP: Unifying Graph Pre-training with Multi-task Graph Dual Prompt [67.8934749027315]
We propose a unified framework for graph hybrid pre-training which injects the task identification and position identification into GNNs.
We also propose a novel pre-training paradigm based on a group of $k$-nearest neighbors.
arXiv Detail & Related papers (2023-10-23T12:11:13Z) - Network Alignment with Transferable Graph Autoencoders [79.89704126746204]
We propose a novel graph autoencoder architecture designed to extract powerful and robust node embeddings.
We prove that the generated embeddings are associated with the eigenvalues and eigenvectors of the graphs.
Our proposed framework also leverages transfer learning and data augmentation to achieve efficient network alignment at a very large scale without retraining.
arXiv Detail & Related papers (2023-10-05T02:58:29Z) - DEGREE: Decomposition Based Explanation For Graph Neural Networks [55.38873296761104]
We propose DEGREE to provide a faithful explanation for GNN predictions.
By decomposing the information generation and aggregation mechanism of GNNs, DEGREE allows tracking the contributions of specific components of the input graph to the final prediction.
We also design a subgraph level interpretation algorithm to reveal complex interactions between graph nodes that are overlooked by previous methods.
arXiv Detail & Related papers (2023-05-22T10:29:52Z) - Deep Graph Reprogramming [112.34663053130073]
"Deep graph reprogramming" is a model reusing task tailored for graph neural networks (GNNs)
We propose an innovative Data Reprogramming paradigm alongside a Model Reprogramming paradigm.
arXiv Detail & Related papers (2023-04-28T02:04:29Z) - Knowledge Graph Completion with Pre-trained Multimodal Transformer and
Twins Negative Sampling [13.016173217017597]
We propose a VisualBERT-enhanced Knowledge Graph Completion model (VBKGC) for short.
VBKGC could capture deeply fused multimodal information for entities and integrate them into the KGC model.
We conduct extensive experiments to show the outstanding performance of VBKGC on the link prediction task.
arXiv Detail & Related papers (2022-09-15T06:50:31Z) - Single-Layer Graph Convolutional Networks For Recommendation [17.3621098912528]
Graph Convolutional Networks (GCNs) have received significant attention and achieved start-of-the-art performances on recommendation tasks.
Existing GCN models tend to perform recursion aggregations among all related nodes, which arises severe computational burden.
We propose a single GCN layer to aggregate information from the neighbors filtered by DA similarity and then generates the node representations.
arXiv Detail & Related papers (2020-06-07T14:38:47Z) - A Robust Hierarchical Graph Convolutional Network Model for
Collaborative Filtering [0.0]
Graph Convolutional Network (GCN) has achieved great success and has been applied in various fields including recommender systems.
GCN still suffers from many issues such as training difficulties, over-smoothing, vulnerable to adversarial attacks, etc.
arXiv Detail & Related papers (2020-04-30T12:50:39Z) - Cross-GCN: Enhancing Graph Convolutional Network with $k$-Order Feature
Interactions [153.6357310444093]
Graph Convolutional Network (GCN) is an emerging technique that performs learning and reasoning on graph data.
We argue that existing designs of GCN forgo modeling cross features, making GCN less effective for tasks or data where cross features are important.
We design a new operator named Cross-feature Graph Convolution, which explicitly models the arbitrary-order cross features with complexity linear to feature dimension and order size.
arXiv Detail & Related papers (2020-03-05T13:05:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.