Two Training Strategies for Improving Relation Extraction over Universal
Graph
- URL: http://arxiv.org/abs/2102.06540v1
- Date: Fri, 12 Feb 2021 14:09:35 GMT
- Title: Two Training Strategies for Improving Relation Extraction over Universal
Graph
- Authors: Qin Dai, Naoya Inoue, Ryo Takahashi and Kentaro Inui
- Abstract summary: This paper explores how the Distantly Supervised Relation Extraction (DS-RE) can benefit from the use of a Universal Graph (UG) and a Knowledge Graph (KG)
We first report that this degradation is associated with the difficulty in learning a UG and then propose two training strategies.
Experimental results on both biomedical and NYT10 datasets prove the robustness of our methods and achieve a new state-of-the-art result on the NYT10 dataset.
- Score: 36.06238013119114
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper explores how the Distantly Supervised Relation Extraction (DS-RE)
can benefit from the use of a Universal Graph (UG), the combination of a
Knowledge Graph (KG) and a large-scale text collection. A straightforward
extension of a current state-of-the-art neural model for DS-RE with a UG may
lead to degradation in performance. We first report that this degradation is
associated with the difficulty in learning a UG and then propose two training
strategies: (1) Path Type Adaptive Pretraining, which sequentially trains the
model with different types of UG paths so as to prevent the reliance on a
single type of UG path; and (2) Complexity Ranking Guided Attention mechanism,
which restricts the attention span according to the complexity of a UG path so
as to force the model to extract features not only from simple UG paths but
also from complex ones. Experimental results on both biomedical and NYT10
datasets prove the robustness of our methods and achieve a new state-of-the-art
result on the NYT10 dataset. The code and datasets used in this paper are
available at https://github.com/baodaiqin/UGDSRE.
Related papers
- Enhanced Expressivity in Graph Neural Networks with Lanczos-Based Linear Constraints [7.605749412696919]
Graph Neural Networks (GNNs) excel in handling graph-structured data but often underperform in link prediction tasks.
We present a novel method to enhance the expressivity of GNNs by embedding induced subgraphs into the graph Laplacian matrix's eigenbasis.
Our method achieves 20x and 10x speedup by only requiring 5% and 10% data from the PubMed and OGBL-Vessel datasets.
arXiv Detail & Related papers (2024-08-22T12:22:00Z) - Can Modifying Data Address Graph Domain Adaptation? [20.343259091425708]
Unsupervised Graph Domain Adaptation (UGDA) aims to facilitate knowledge transfer from a labeled source graph to an unlabeled target graph.
We propose GraphAlign, a novel UGDA method that generates a small yet transferable graph.
By exclusively training a GNN on this new graph with classic Empirical Risk Minimization (ERM), GraphAlign attains exceptional performance on the target graph.
arXiv Detail & Related papers (2024-07-27T17:56:31Z) - Revisiting, Benchmarking and Understanding Unsupervised Graph Domain Adaptation [31.106636947179005]
Unsupervised Graph Domain Adaptation involves the transfer of knowledge from a label-rich source graph to an unlabeled target graph.
We present the first comprehensive benchmark for unsupervised graph domain adaptation named GDABench.
We observe that the performance of current UGDA models varies significantly across different datasets and adaptation scenarios.
arXiv Detail & Related papers (2024-07-09T06:44:09Z) - Efficient Heterogeneous Graph Learning via Random Projection [58.4138636866903]
Heterogeneous Graph Neural Networks (HGNNs) are powerful tools for deep learning on heterogeneous graphs.
Recent pre-computation-based HGNNs use one-time message passing to transform a heterogeneous graph into regular-shaped tensors.
We propose a hybrid pre-computation-based HGNN, named Random Projection Heterogeneous Graph Neural Network (RpHGNN)
arXiv Detail & Related papers (2023-10-23T01:25:44Z) - Label Deconvolution for Node Representation Learning on Large-scale
Attributed Graphs against Learning Bias [75.44877675117749]
We propose an efficient label regularization technique, namely Label Deconvolution (LD), to alleviate the learning bias by a novel and highly scalable approximation to the inverse mapping of GNNs.
Experiments demonstrate LD significantly outperforms state-of-the-art methods on Open Graph datasets Benchmark.
arXiv Detail & Related papers (2023-09-26T13:09:43Z) - LUKE-Graph: A Transformer-based Approach with Gated Relational Graph
Attention for Cloze-style Reading Comprehension [13.173307471333619]
We propose the LUKE-Graph, a model that builds a heterogeneous graph based on the intuitive relationships between entities in a document.
We then use the Attention reading (RGAT) to fuse the graph's reasoning information and the contextual representation encoded by the pre-trained LUKE model.
Experimental results demonstrate that the LUKE-Graph achieves state-of-the-art performance with commonsense reasoning.
arXiv Detail & Related papers (2023-03-12T14:31:44Z) - Causal Incremental Graph Convolution for Recommender System Retraining [89.25922726558875]
Real-world recommender system needs to be regularly retrained to keep with the new data.
In this work, we consider how to efficiently retrain graph convolution network (GCN) based recommender models.
arXiv Detail & Related papers (2021-08-16T04:20:09Z) - Robust Optimization as Data Augmentation for Large-scale Graphs [117.2376815614148]
We propose FLAG (Free Large-scale Adversarial Augmentation on Graphs), which iteratively augments node features with gradient-based adversarial perturbations during training.
FLAG is a general-purpose approach for graph data, which universally works in node classification, link prediction, and graph classification tasks.
arXiv Detail & Related papers (2020-10-19T21:51:47Z) - Revisiting Graph based Collaborative Filtering: A Linear Residual Graph
Convolutional Network Approach [55.44107800525776]
Graph Convolutional Networks (GCNs) are state-of-the-art graph based representation learning models.
In this paper, we revisit GCN based Collaborative Filtering (CF) based Recommender Systems (RS)
We show that removing non-linearities would enhance recommendation performance, consistent with the theories in simple graph convolutional networks.
We propose a residual network structure that is specifically designed for CF with user-item interaction modeling.
arXiv Detail & Related papers (2020-01-28T04:41:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.