Learning Relation Prototype from Unlabeled Texts for Long-tail Relation
Extraction
- URL: http://arxiv.org/abs/2011.13574v1
- Date: Fri, 27 Nov 2020 06:21:12 GMT
- Title: Learning Relation Prototype from Unlabeled Texts for Long-tail Relation
Extraction
- Authors: Yixin Cao, Jun Kuang, Ming Gao, Aoying Zhou, Yonggang Wen, Tat-Seng
Chua
- Abstract summary: We propose a general approach to learn relation prototypes from unlabeled texts.
We learn relation prototypes as an implicit factor between entities.
We conduct experiments on two publicly available datasets: New York Times and Google Distant Supervision.
- Score: 84.64435075778988
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Relation Extraction (RE) is a vital step to complete Knowledge Graph (KG) by
extracting entity relations from texts.However, it usually suffers from the
long-tail issue. The training data mainly concentrates on a few types of
relations, leading to the lackof sufficient annotations for the remaining types
of relations. In this paper, we propose a general approach to learn relation
prototypesfrom unlabeled texts, to facilitate the long-tail relation extraction
by transferring knowledge from the relation types with sufficient trainingdata.
We learn relation prototypes as an implicit factor between entities, which
reflects the meanings of relations as well as theirproximities for transfer
learning. Specifically, we construct a co-occurrence graph from texts, and
capture both first-order andsecond-order entity proximities for embedding
learning. Based on this, we further optimize the distance from entity pairs
tocorresponding prototypes, which can be easily adapted to almost arbitrary RE
frameworks. Thus, the learning of infrequent or evenunseen relation types will
benefit from semantically proximate relations through pairs of entities and
large-scale textual information.We have conducted extensive experiments on two
publicly available datasets: New York Times and Google Distant
Supervision.Compared with eight state-of-the-art baselines, our proposed model
achieves significant improvements (4.1% F1 on average). Furtherresults on
long-tail relations demonstrate the effectiveness of the learned relation
prototypes. We further conduct an ablation study toinvestigate the impacts of
varying components, and apply it to four basic relation extraction models to
verify the generalization ability.Finally, we analyze several example cases to
give intuitive impressions as qualitative analysis. Our codes will be released
later.
Related papers
- Cross-Domain Few-Shot Relation Extraction via Representation Learning
and Domain Adaptation [1.1602089225841632]
Few-shot relation extraction aims to recognize novel relations with few labeled sentences in each relation.
Previous metric-based few-shot relation extraction algorithms identify relationships by comparing the prototypes generated by the few labeled sentences embedding with the embeddings of the query sentences using a trained metric function.
We suggest learning more interpretable and efficient prototypes from prior knowledge and the intrinsic semantics of relations to extract new relations in various domains more effectively.
arXiv Detail & Related papers (2022-12-05T19:34:52Z) - A Simple yet Effective Relation Information Guided Approach for Few-Shot
Relation Extraction [22.60428265210431]
Few-Shot Relation Extraction aims at predicting the relation for a pair of entities in a sentence by training with a few labelled examples in each relation.
Some recent works have introduced relation information to assist model learning based on Prototype Network.
We argue that relation information can be introduced more explicitly and effectively into the model.
arXiv Detail & Related papers (2022-05-19T13:03:01Z) - Prototypical Representation Learning for Relation Extraction [56.501332067073065]
This paper aims to learn predictive, interpretable, and robust relation representations from distantly-labeled data.
We learn prototypes for each relation from contextual information to best explore the intrinsic semantics of relations.
Results on several relation learning tasks show that our model significantly outperforms the previous state-of-the-art relational models.
arXiv Detail & Related papers (2021-03-22T08:11:43Z) - Learning Informative Representations of Biomedical Relations with Latent
Variable Models [2.4366811507669115]
We propose a latent variable model with an arbitrarily flexible distribution to represent the relation between an entity pair.
We demonstrate that our model achieves results competitive with strong baselines for both tasks while having fewer parameters and being significantly faster to train.
arXiv Detail & Related papers (2020-11-20T08:56:31Z) - Bridging Text and Knowledge with Multi-Prototype Embedding for Few-Shot
Relational Triple Extraction [40.00702385889112]
We propose a novel multi-prototype embedding network model to jointly extract the composition of relational triples.
We design a hybrid learning mechanism that bridges text and knowledge concerning both entities and relations.
Experimental results demonstrate that the proposed method can improve the performance of the few-shot triple extraction.
arXiv Detail & Related papers (2020-10-30T04:18:39Z) - Learning to Decouple Relations: Few-Shot Relation Classification with
Entity-Guided Attention and Confusion-Aware Training [49.9995628166064]
We propose CTEG, a model equipped with two mechanisms to learn to decouple easily-confused relations.
On the one hand, an EGA mechanism is introduced to guide the attention to filter out information causing confusion.
On the other hand, a Confusion-Aware Training (CAT) method is proposed to explicitly learn to distinguish relations.
arXiv Detail & Related papers (2020-10-21T11:07:53Z) - Improving Long-Tail Relation Extraction with Collaborating
Relation-Augmented Attention [63.26288066935098]
We propose a novel neural network, Collaborating Relation-augmented Attention (CoRA), to handle both the wrong labeling and long-tail relations.
In the experiments on the popular benchmark dataset NYT, the proposed CoRA improves the prior state-of-the-art performance by a large margin.
arXiv Detail & Related papers (2020-10-08T05:34:43Z) - Type-augmented Relation Prediction in Knowledge Graphs [65.88395564516115]
We propose a type-augmented relation prediction (TaRP) method, where we apply both the type information and instance-level information for relation prediction.
Our proposed TaRP method achieves significantly better performance than state-of-the-art methods on four benchmark datasets.
arXiv Detail & Related papers (2020-09-16T21:14:18Z) - Relation-Guided Representation Learning [53.60351496449232]
We propose a new representation learning method that explicitly models and leverages sample relations.
Our framework well preserves the relations between samples.
By seeking to embed samples into subspace, we show that our method can address the large-scale and out-of-sample problem.
arXiv Detail & Related papers (2020-07-11T10:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.