Relation-Specific Attentions over Entity Mentions for Enhanced
Document-Level Relation Extraction
- URL: http://arxiv.org/abs/2205.14393v1
- Date: Sat, 28 May 2022 10:40:31 GMT
- Title: Relation-Specific Attentions over Entity Mentions for Enhanced
Document-Level Relation Extraction
- Authors: Jiaxin Yu and Deqing Yang and Shuyu Tian
- Abstract summary: We propose RSMAN which performs selective attentions over different entity mentions with respect to candidate relations.
Our experiments upon two benchmark datasets show that our RSMAN can bring significant improvements for some backbone models.
- Score: 4.685620089585031
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Compared with traditional sentence-level relation extraction, document-level
relation extraction is a more challenging task where an entity in a document
may be mentioned multiple times and associated with multiple relations.
However, most methods of document-level relation extraction do not distinguish
between mention-level features and entity-level features, and just apply simple
pooling operation for aggregating mention-level features into entity-level
features. As a result, the distinct semantics between the different mentions of
an entity are overlooked. To address this problem, we propose RSMAN in this
paper which performs selective attentions over different entity mentions with
respect to candidate relations. In this manner, the flexible and
relation-specific representations of entities are obtained which indeed benefit
relation classification. Our extensive experiments upon two benchmark datasets
show that our RSMAN can bring significant improvements for some backbone models
to achieve state-of-the-art performance, especially when an entity have
multiple mentions in the document.
Related papers
- Generative Retrieval Meets Multi-Graded Relevance [104.75244721442756]
We introduce a framework called GRaded Generative Retrieval (GR$2$)
GR$2$ focuses on two key components: ensuring relevant and distinct identifiers, and implementing multi-graded constrained contrastive training.
Experiments on datasets with both multi-graded and binary relevance demonstrate the effectiveness of GR$2$.
arXiv Detail & Related papers (2024-09-27T02:55:53Z) - Hypergraph based Understanding for Document Semantic Entity Recognition [65.84258776834524]
We build a novel hypergraph attention document semantic entity recognition framework, HGA, which uses hypergraph attention to focus on entity boundaries and entity categories at the same time.
Our results on FUNSD, CORD, XFUNDIE show that our method can effectively improve the performance of semantic entity recognition tasks.
arXiv Detail & Related papers (2024-07-09T14:35:49Z) - Entity Disambiguation via Fusion Entity Decoding [68.77265315142296]
We propose an encoder-decoder model to disambiguate entities with more detailed entity descriptions.
We observe +1.5% improvements in end-to-end entity linking in the GERBIL benchmark compared with EntQA.
arXiv Detail & Related papers (2024-04-02T04:27:54Z) - DREQ: Document Re-Ranking Using Entity-based Query Understanding [6.675805308519988]
DREQ is an entity-oriented dense document re-ranking model.
We emphasize the query-relevant entities within a document's representation while simultaneously attenuating the less relevant ones.
We show that DREQ outperforms state-of-the-art neural and non-neural re-ranking methods.
arXiv Detail & Related papers (2024-01-11T14:27:12Z) - Anaphor Assisted Document-Level Relation Extraction [42.57958231709678]
Document-level relation extraction (DocRE) involves identifying relations between entities distributed in multiple sentences within a document.
Existing methods focus on building a heterogeneous document graph to model the internal structure of an entity and the external interaction between entities.
We propose an Anaphor-Assisted (AA) framework for DocRE tasks.
arXiv Detail & Related papers (2023-10-28T06:11:18Z) - Mutually Guided Few-shot Learning for Relational Triple Extraction [10.539566491939844]
Mutually Guided Few-shot learning framework for Triple Extraction (MG-FTE)
Our method consists of an entity-guided relation-decoder to classify relations and a proto-decoder to extract entities.
Our method outperforms many state-of-the-art methods by 12.6 F1 score on FewRel 1.0 (single domain) and 20.5 F1 score on FewRel 2.0 (cross-domain)
arXiv Detail & Related papers (2023-06-23T06:15:54Z) - HIORE: Leveraging High-order Interactions for Unified Entity Relation
Extraction [85.80317530027212]
We propose HIORE, a new method for unified entity relation extraction.
The key insight is to leverage the complex association among word pairs, which contains richer information than the first-order word-by-word interactions.
Experiments show that HIORE achieves the state-of-the-art performance on relation extraction and an improvement of 1.11.8 F1 points over the prior best unified model.
arXiv Detail & Related papers (2023-05-07T14:57:42Z) - ReSel: N-ary Relation Extraction from Scientific Text and Tables by
Learning to Retrieve and Select [53.071352033539526]
We study the problem of extracting N-ary relations from scientific articles.
Our proposed method ReSel decomposes this task into a two-stage procedure.
Our experiments on three scientific information extraction datasets show that ReSel outperforms state-of-the-art baselines significantly.
arXiv Detail & Related papers (2022-10-26T02:28:02Z) - Entity Relation Extraction as Dependency Parsing in Visually Rich
Documents [18.67730663266417]
We adapt the popular dependency parsing model, the biaffine, to this entity relation extraction task.
Being different from the original dependency parsing model which recognizes dependency relations between words, we identify relations between groups of words with layout information instead.
As for the real-world application, our model has been applied to the in-house customs data, achieving reliable performance in the production setting.
arXiv Detail & Related papers (2021-10-19T12:26:40Z) - Mention-centered Graph Neural Network for Document-level Relation
Extraction [2.724649366608364]
We build cross-sentence dependencies by inferring compositional relations between inter-sentence mentions.
Experiments show the connections between different mentions are crucial to document-level relation extraction.
arXiv Detail & Related papers (2021-03-15T08:19:44Z) - Learning Relation Prototype from Unlabeled Texts for Long-tail Relation
Extraction [84.64435075778988]
We propose a general approach to learn relation prototypes from unlabeled texts.
We learn relation prototypes as an implicit factor between entities.
We conduct experiments on two publicly available datasets: New York Times and Google Distant Supervision.
arXiv Detail & Related papers (2020-11-27T06:21:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.