A shallow neural model for relation prediction
- URL: http://arxiv.org/abs/2101.09090v1
- Date: Fri, 22 Jan 2021 13:10:11 GMT
- Title: A shallow neural model for relation prediction
- Authors: Caglar Demir and Diego Moussallem and Axel-Cyrille Ngonga Ngomo
- Abstract summary: We propose a shallow neural model (SHALLOM) that accurately infers missing relations from entities.
Our experiments indicate that SHALLOM outperforms state-of-the-art approaches on the FB15K-237 and WN18RR datasets.
- Score: 2.2559617939136505
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Knowledge graph completion refers to predicting missing triples. Most
approaches achieve this goal by predicting entities, given an entity and a
relation. We predict missing triples via the relation prediction. To this end,
we frame the relation prediction problem as a multi-label classification
problem and propose a shallow neural model (SHALLOM) that accurately infers
missing relations from entities. SHALLOM is analogous to C-BOW as both
approaches predict a central token (p) given surrounding tokens ((s,o)). Our
experiments indicate that SHALLOM outperforms state-of-the-art approaches on
the FB15K-237 and WN18RR with margins of up to $3\%$ and $8\%$ (absolute),
respectively, while requiring a maximum training time of 8 minutes on these
datasets. We ensure the reproducibility of our results by providing an
open-source implementation including training and evaluation scripts at
{\url{https://github.com/dice-group/Shallom}.}
Related papers
- TokenUnify: Scalable Autoregressive Visual Pre-training with Mixture Token Prediction [61.295716741720284]
TokenUnify is a novel pretraining method that integrates random token prediction, next-token prediction, and next-all token prediction.
Cooperated with TokenUnify, we have assembled a large-scale electron microscopy (EM) image dataset with ultra-high resolution.
This dataset includes over 120 million annotated voxels, making it the largest neuron segmentation dataset to date.
arXiv Detail & Related papers (2024-05-27T05:45:51Z) - Joint Prediction Regions for time-series models [0.0]
It is an easy task to compute Joint Prediction regions (JPR) when the data is IID.
This project aims to implement Wolf and Wunderli's method for constructing JPRs and compare it with other methods.
arXiv Detail & Related papers (2024-05-14T02:38:49Z) - Probabilistic Demand Forecasting with Graph Neural Networks [0.0]
This paper builds on previous research on Graph Neural Networks (GNNs) and makes two contributions.
First, we integrate a GNN encoder into a state-of-the-art DeepAR model. The combined model produces probabilistic forecasts, which are crucial for decision-making under uncertainty.
Second, we propose to build graphs using article similarity, which avoids reliance on a pre-defined graph structure. Experiments on three real-world datasets show that the proposed approach consistently outperforms non-graph benchmarks.
arXiv Detail & Related papers (2024-01-23T21:20:48Z) - You Only Forward Once: Prediction and Rationalization in A Single
Forward Pass [10.998983921416533]
Unsupervised rationale extraction aims to extract concise and contiguous text snippets to support model predictions without any rationale.
Previous studies have used a two-phase framework known as the Rationalizing Neural Prediction (RNP) framework, which follows a generate-then-predict paradigm.
We propose a novel single-phase framework called You Only Forward Once (YOFO), derived from a relaxed version of rationale where rationales aim to support model predictions rather than make predictions.
arXiv Detail & Related papers (2023-11-04T08:04:28Z) - FABind: Fast and Accurate Protein-Ligand Binding [127.7790493202716]
$mathbfFABind$ is an end-to-end model that combines pocket prediction and docking to achieve accurate and fast protein-ligand binding.
Our proposed model demonstrates strong advantages in terms of effectiveness and efficiency compared to existing methods.
arXiv Detail & Related papers (2023-10-10T16:39:47Z) - Normalizing Flow-based Neural Process for Few-Shot Knowledge Graph
Completion [69.55700751102376]
Few-shot knowledge graph completion (FKGC) aims to predict missing facts for unseen relations with few-shot associated facts.
Existing FKGC methods are based on metric learning or meta-learning, which often suffer from the out-of-distribution and overfitting problems.
In this paper, we propose a normalizing flow-based neural process for few-shot knowledge graph completion (NP-FKGC)
arXiv Detail & Related papers (2023-04-17T11:42:28Z) - Convolutional Hypercomplex Embeddings for Link Prediction [2.6209112069534046]
We propose QMult, OMult, ConvQ and ConvO to tackle the link prediction problem.
QMult, OMult, ConvQ and ConvO build upon QMult and OMult by including convolution operations in a way inspired by the residual learning framework.
We evaluate our approaches on seven link prediction datasets including WN18RR, FB15K-237 and YAGO3-10.
arXiv Detail & Related papers (2021-06-29T10:26:51Z) - Injecting Knowledge in Data-driven Vehicle Trajectory Predictors [82.91398970736391]
Vehicle trajectory prediction tasks have been commonly tackled from two perspectives: knowledge-driven or data-driven.
In this paper, we propose to learn a "Realistic Residual Block" (RRB) which effectively connects these two perspectives.
Our proposed method outputs realistic predictions by confining the residual range and taking into account its uncertainty.
arXiv Detail & Related papers (2021-03-08T16:03:09Z) - Knowledge Generation -- Variational Bayes on Knowledge Graphs [0.685316573653194]
This thesis is a proof of concept for potential of Vari Auto-Encoder (VAE) on representation of real-world Knowledge Graphs.
Inspired by successful approaches to generation graphs, we evaluate the capabilities of our model, the Variational Auto-Encoder (RGVAE)
The RGVAE is first evaluated on link prediction. The mean reciprocal rank (MRR) scores on the two FB15K-237 and WN18RR datasets are compared.
We investigate the latent space in a twofold experiment: first, linear between the latent representation of two triples, then the exploration of each
arXiv Detail & Related papers (2021-01-21T21:23:17Z) - Evaluating Prediction-Time Batch Normalization for Robustness under
Covariate Shift [81.74795324629712]
We call prediction-time batch normalization, which significantly improves model accuracy and calibration under covariate shift.
We show that prediction-time batch normalization provides complementary benefits to existing state-of-the-art approaches for improving robustness.
The method has mixed results when used alongside pre-training, and does not seem to perform as well under more natural types of dataset shift.
arXiv Detail & Related papers (2020-06-19T05:08:43Z) - Parameter Space Factorization for Zero-Shot Learning across Tasks and
Languages [112.65994041398481]
We propose a Bayesian generative model for the space of neural parameters.
We infer the posteriors over such latent variables based on data from seen task-language combinations.
Our model yields comparable or better results than state-of-the-art, zero-shot cross-lingual transfer methods.
arXiv Detail & Related papers (2020-01-30T16:58:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.