Multi-Aspect Explainable Inductive Relation Prediction by Sentence
Transformer
- URL: http://arxiv.org/abs/2301.01664v2
- Date: Mon, 1 May 2023 09:52:04 GMT
- Title: Multi-Aspect Explainable Inductive Relation Prediction by Sentence
Transformer
- Authors: Zhixiang Su, Di Wang, Chunyan Miao, Lizhen Cui
- Abstract summary: We introduce the concepts of relation path coverage and relation path confidence to filter out unreliable paths prior to model training to elevate the model performance.
We propose Knowledge Reasoning Sentence Transformer (KRST) to predict inductive relations in knowledge graphs.
- Score: 60.75757851637566
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent studies on knowledge graphs (KGs) show that path-based methods
empowered by pre-trained language models perform well in the provision of
inductive and explainable relation predictions. In this paper, we introduce the
concepts of relation path coverage and relation path confidence to filter out
unreliable paths prior to model training to elevate the model performance.
Moreover, we propose Knowledge Reasoning Sentence Transformer (KRST) to predict
inductive relations in KGs. KRST is designed to encode the extracted reliable
paths in KGs, allowing us to properly cluster paths and provide multi-aspect
explanations. We conduct extensive experiments on three real-world datasets.
The experimental results show that compared to SOTA models, KRST achieves the
best performance in most transductive and inductive test cases (4 of 6), and in
11 of 12 few-shot test cases.
Related papers
- Training Nonlinear Transformers for Chain-of-Thought Inference: A Theoretical Generalization Analysis [82.51626700527837]
Chain-of-shift (CoT) is an efficient method that enables the reasoning ability of large language models by augmenting the query using examples with multiple intermediate steps.
We show that despite the theoretical success of CoT, it fails to provide an accurate generalization when CoT does.
arXiv Detail & Related papers (2024-10-03T03:12:51Z) - Path-based Explanation for Knowledge Graph Completion [17.541247786437484]
Proper explanations for the results of GNN-based Knowledge Graph Completion models increase model transparency.
Existing practices for explaining KGC tasks rely on instance/subgraph-based approaches.
We propose Power-Link, the first path-based KGC explainer that explores GNN-based models.
arXiv Detail & Related papers (2024-01-04T14:19:37Z) - Towards Faithful Explanations for Text Classification with Robustness
Improvement and Explanation Guided Training [30.626080706755822]
Feature attribution methods highlight the important input tokens as explanations to model predictions.
Recent works show that explanations provided by these methods face challenges of being faithful and robust.
We propose a method with Robustness improvement and Explanation Guided training towards more faithful EXplanations (REGEX) for text classification.
arXiv Detail & Related papers (2023-12-29T13:07:07Z) - Anchoring Path for Inductive Relation Prediction in Knowledge Graphs [69.81600732388182]
APST takes both APs and CPs as the inputs of a unified Sentence Transformer architecture.
We evaluate APST on three public datasets and achieve state-of-the-art (SOTA) performance in 30 of 36 transductive, inductive, and few-shot experimental settings.
arXiv Detail & Related papers (2023-12-21T06:02:25Z) - KGEx: Explaining Knowledge Graph Embeddings via Subgraph Sampling and
Knowledge Distillation [6.332573781489264]
We present KGEx, a novel method that explains individual link predictions by drawing inspiration from surrogate models research.
Given a target triple to predict, KGEx trains surrogate KGE models that we use to identify important training triples.
We conduct extensive experiments on two publicly available datasets, to demonstrate that KGEx is capable of providing explanations faithful to the black-box model.
arXiv Detail & Related papers (2023-10-02T10:20:24Z) - Towards Few-shot Inductive Link Prediction on Knowledge Graphs: A
Relational Anonymous Walk-guided Neural Process Approach [49.00753238429618]
Few-shot inductive link prediction on knowledge graphs aims to predict missing links for unseen entities with few-shot links observed.
Recent inductive methods utilize the sub-graphs around unseen entities to obtain the semantics and predict links inductively.
We propose a novel relational anonymous walk-guided neural process for few-shot inductive link prediction on knowledge graphs, denoted as RawNP.
arXiv Detail & Related papers (2023-06-26T12:02:32Z) - CASTLE: Regularization via Auxiliary Causal Graph Discovery [89.74800176981842]
We introduce Causal Structure Learning (CASTLE) regularization and propose to regularize a neural network by jointly learning the causal relationships between variables.
CASTLE efficiently reconstructs only the features in the causal DAG that have a causal neighbor, whereas reconstruction-based regularizers suboptimally reconstruct all input features.
arXiv Detail & Related papers (2020-09-28T09:49:38Z) - Learning Reasoning Strategies in End-to-End Differentiable Proving [50.9791149533921]
Conditional Theorem Provers learn optimal rule selection strategy via gradient-based optimisation.
We show that Conditional Theorem Provers are scalable and yield state-of-the-art results on the CLUTRR dataset.
arXiv Detail & Related papers (2020-07-13T16:22:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.