HyperMono: A Monotonicity-aware Approach to Hyper-Relational Knowledge Representation
- URL: http://arxiv.org/abs/2404.09848v2
- Date: Tue, 13 Aug 2024 09:51:39 GMT
- Title: HyperMono: A Monotonicity-aware Approach to Hyper-Relational Knowledge Representation
- Authors: Zhiwei Hu, Víctor Gutiérrez-Basulto, Zhiliang Xiang, Ru Li, Jeff Z. Pan,
- Abstract summary: In a hyper-relational knowledge graph (HKG), each fact is composed of a main triple associated with attribute-value qualifiers, which express additional factual knowledge.
This paper proposes the HyperMono model for hyper-relational knowledge graph completion, which realizes stage reasoning and qualifier monotonicity.
- Score: 27.28214706269035
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In a hyper-relational knowledge graph (HKG), each fact is composed of a main triple associated with attribute-value qualifiers, which express additional factual knowledge. The hyper-relational knowledge graph completion (HKGC) task aims at inferring plausible missing links in a HKG. Most existing approaches to HKGC focus on enhancing the communication between qualifier pairs and main triples, while overlooking two important properties that emerge from the monotonicity of the hyper-relational graphs representation regime. Stage Reasoning allows for a two-step reasoning process, facilitating the integration of coarse-grained inference results derived solely from main triples and fine-grained inference results obtained from hyper-relational facts with qualifiers. In the initial stage, coarse-grained results provide an upper bound for correct predictions, which are subsequently refined in the fine-grained step. More generally, Qualifier Monotonicity implies that by attaching more qualifier pairs to a main triple, we may only narrow down the answer set, but never enlarge it. This paper proposes the HyperMono model for hyper-relational knowledge graph completion, which realizes stage reasoning and qualifier monotonicity. To implement qualifier monotonicity HyperMono resorts to cone embeddings. Experiments on three real-world datasets with three different scenario conditions demonstrate the strong performance of HyperMono when compared to the SoTA.
Related papers
- Hypergraph Transformer for Semi-Supervised Classification [50.92027313775934]
We propose a novel hypergraph learning framework, HyperGraph Transformer (HyperGT)
HyperGT uses a Transformer-based neural network architecture to effectively consider global correlations among all nodes and hyperedges.
It achieves comprehensive hypergraph representation learning by effectively incorporating global interactions while preserving local connectivity patterns.
arXiv Detail & Related papers (2023-12-18T17:50:52Z) - Enhancing Hyperedge Prediction with Context-Aware Self-Supervised
Learning [64.46188414653204]
We propose a novel hyperedge prediction framework (CASH)
CASH employs context-aware node aggregation to capture complex relations among nodes in each hyperedge for (C1) and (2) self-supervised contrastive learning in the context of hyperedge prediction to enhance hypergraph representations for (C2)
Experiments on six real-world hypergraphs reveal that CASH consistently outperforms all competing methods in terms of the accuracy in hyperedge prediction.
arXiv Detail & Related papers (2023-09-11T20:06:00Z) - HyperFormer: Enhancing Entity and Relation Interaction for
Hyper-Relational Knowledge Graph Completion [25.399684403558553]
Hyper-relational knowledge graphs (HKGs) extend standard knowledge graphs by associating attribute-value qualifiers to triples.
We propose HyperFormer, a model that considers local-level sequential information, which encodes the content of the entities, relations and qualifiers of a triple.
arXiv Detail & Related papers (2023-08-12T09:31:43Z) - Shrinking Embeddings for Hyper-Relational Knowledge Graphs [42.23862602535092]
We present emphShrinkE, a geometric hyper-relational KG embedding method aiming to explicitly model these patterns.
Experimental results demonstrate ShrinkE's superiority on three benchmarks of hyper-relational KGs.
arXiv Detail & Related papers (2023-06-03T21:14:59Z) - Few-shot Link Prediction on N-ary Facts [70.8150181683017]
Link Prediction on Hyper-relational Facts (LPHFs) is to predict a missing element in a hyper-relational fact.
Few-Shot Link Prediction on Hyper-relational Facts (PHFs) aims to predict a missing entity in a hyper-relational fact with limited support instances.
arXiv Detail & Related papers (2023-05-10T12:44:00Z) - A Dataset for Hyper-Relational Extraction and a Cube-Filling Approach [59.89749342550104]
We propose the task of hyper-relational extraction to extract more specific and complete facts from text.
Existing models cannot perform hyper-relational extraction as it requires a model to consider the interaction between three entities.
We propose CubeRE, a cube-filling model inspired by table-filling approaches and explicitly considers the interaction between relation triplets and qualifiers.
arXiv Detail & Related papers (2022-11-18T03:51:28Z) - Learning Representations for Hyper-Relational Knowledge Graphs [35.380689788802776]
We design a framework to learn representations for hyper-relational facts using multiple aggregators.
Experiments demonstrate the effectiveness of our framework across multiple datasets.
We conduct an ablation study that validates the importance of the various components in our framework.
arXiv Detail & Related papers (2022-08-30T15:02:14Z) - ExpressivE: A Spatio-Functional Embedding For Knowledge Graph Completion [78.8942067357231]
ExpressivE embeds pairs of entities as points and relations as hyper-parallelograms in the virtual triple space.
We show that ExpressivE is competitive with state-of-the-art KGEs and even significantly outperforms them on W18RR.
arXiv Detail & Related papers (2022-06-08T23:34:39Z) - Message Passing for Hyper-Relational Knowledge Graphs [7.733963597282456]
We propose a message passing graph encoder - StarE capable of modeling such hyper-relational knowledge graphs.
StarE can encode an arbitrary number of additional information (qualifiers) along with the main triple while keeping the semantic roles of qualifiers and triples intact.
Our experiments demonstrate that StarE based LP model outperforms existing approaches across multiple benchmarks.
arXiv Detail & Related papers (2020-09-22T22:38:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.