Multi-Scale Feature and Metric Learning for Relation Extraction
- URL: http://arxiv.org/abs/2107.13425v1
- Date: Wed, 28 Jul 2021 15:14:36 GMT
- Title: Multi-Scale Feature and Metric Learning for Relation Extraction
- Authors: Mi Zhang, Tieyun Qian
- Abstract summary: We propose a multi-scale feature and metric learning framework for relation extraction.
Specifically, we first develop a multi-scale convolutional neural network to aggregate the non-successive mainstays in the lexical sequence.
We also design a multi-scale graph convolutional network which can increase the receptive field towards specific syntactic roles.
- Score: 15.045539169021092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing methods in relation extraction have leveraged the lexical features
in the word sequence and the syntactic features in the parse tree. Though
effective, the lexical features extracted from the successive word sequence may
introduce some noise that has little or no meaningful content. Meanwhile, the
syntactic features are usually encoded via graph convolutional networks which
have restricted receptive field. To address the above limitations, we propose a
multi-scale feature and metric learning framework for relation extraction.
Specifically, we first develop a multi-scale convolutional neural network to
aggregate the non-successive mainstays in the lexical sequence. We also design
a multi-scale graph convolutional network which can increase the receptive
field towards specific syntactic roles. Moreover, we present a multi-scale
metric learning paradigm to exploit both the feature-level relation between
lexical and syntactic features and the sample-level relation between instances
with the same or different classes. We conduct extensive experiments on three
real world datasets for various types of relation extraction tasks. The results
demonstrate that our model significantly outperforms the state-of-the-art
approaches.
Related papers
- GSSF: Generalized Structural Sparse Function for Deep Cross-modal Metric Learning [51.677086019209554]
We propose a Generalized Structural Sparse to capture powerful relationships across modalities for pair-wise similarity learning.
The distance metric delicately encapsulates two formats of diagonal and block-diagonal terms.
Experiments on cross-modal and two extra uni-modal retrieval tasks have validated its superiority and flexibility.
arXiv Detail & Related papers (2024-10-20T03:45:50Z) - Conversational Semantic Parsing using Dynamic Context Graphs [68.72121830563906]
We consider the task of conversational semantic parsing over general purpose knowledge graphs (KGs) with millions of entities, and thousands of relation-types.
We focus on models which are capable of interactively mapping user utterances into executable logical forms.
arXiv Detail & Related papers (2023-05-04T16:04:41Z) - Syntactic Multi-view Learning for Open Information Extraction [26.1066324477346]
Open Information Extraction (OpenIE) aims to extracts from open-domain sentences.
In this paper, we model both constituency and dependency trees into word-level graphs.
arXiv Detail & Related papers (2022-12-05T07:15:41Z) - Lexical semantics enhanced neural word embeddings [4.040491121427623]
hierarchy-fitting is a novel approach to modelling semantic similarity nuances inherently stored in the IS-A hierarchies.
Results demonstrate the efficacy of hierarchy-fitting in specialising neural embeddings with semantic relations in late fusion.
arXiv Detail & Related papers (2022-10-03T08:10:23Z) - Learning the Evolutionary and Multi-scale Graph Structure for
Multivariate Time Series Forecasting [50.901984244738806]
We show how to model the evolutionary and multi-scale interactions of time series.
In particular, we first provide a hierarchical graph structure cooperated with the dilated convolution to capture the scale-specific correlations.
A unified neural network is provided to integrate the components above to get the final prediction.
arXiv Detail & Related papers (2022-06-28T08:11:12Z) - Modeling Multi-Granularity Hierarchical Features for Relation Extraction [26.852869800344813]
We propose a novel method to extract multi-granularity features based solely on the original input sentences.
We show that effective structured features can be attained even without external knowledge.
arXiv Detail & Related papers (2022-04-09T09:44:05Z) - Multiplex Graph Neural Network for Extractive Text Summarization [34.185093491514394]
Extractive text summarization aims at extracting the most representative sentences from a given document as its summary.
We propose a novel Multiplex Graph Convolutional Network (Multi-GCN) to jointly model different types of relationships among sentences and words.
Based on Multi-GCN, we propose a Multiplex Graph Summarization (Multi-GraS) model for extractive text summarization.
arXiv Detail & Related papers (2021-08-29T16:11:01Z) - Learning the Implicit Semantic Representation on Graph-Structured Data [57.670106959061634]
Existing representation learning methods in graph convolutional networks are mainly designed by describing the neighborhood of each node as a perceptual whole.
We propose a Semantic Graph Convolutional Networks (SGCN) that explores the implicit semantics by learning latent semantic-paths in graphs.
arXiv Detail & Related papers (2021-01-16T16:18:43Z) - Towards Interpretable Multi-Task Learning Using Bilevel Programming [18.293397644865454]
Interpretable Multi-Task Learning can be expressed as learning a sparse graph of the task relationship based on the prediction performance of the learned models.
We show empirically how the induced sparse graph improves the interpretability of the learned models and their relationship on synthetic and real data, without sacrificing generalization performance.
arXiv Detail & Related papers (2020-09-11T15:04:27Z) - Heterogeneous Graph Neural Networks for Extractive Document
Summarization [101.17980994606836]
Cross-sentence relations are a crucial step in extractive document summarization.
We present a graph-based neural network for extractive summarization (HeterSumGraph)
We introduce different types of nodes into graph-based neural networks for extractive document summarization.
arXiv Detail & Related papers (2020-04-26T14:38:11Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.