DORE: Document Ordered Relation Extraction based on Generative Framework
- URL: http://arxiv.org/abs/2210.16064v1
- Date: Fri, 28 Oct 2022 11:18:10 GMT
- Title: DORE: Document Ordered Relation Extraction based on Generative Framework
- Authors: Qipeng Guo, Yuqing Yang, Hang Yan, Xipeng Qiu, Zheng Zhang
- Abstract summary: This paper investigates the root cause of the underwhelming performance of the existing generative DocRE models.
We propose to generate a symbolic and ordered sequence from the relation matrix which is deterministic and easier for model to learn.
Experimental results on four datasets show that our proposed method can improve the performance of the generative DocRE models.
- Score: 56.537386636819626
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, there is a surge of generation-based information extraction
work, which allows a more direct use of pre-trained language models and
efficiently captures output dependencies. However, previous generative methods
using lexical representation do not naturally fit document-level relation
extraction (DocRE) where there are multiple entities and relational facts. In
this paper, we investigate the root cause of the underwhelming performance of
the existing generative DocRE models and discover that the culprit is the
inadequacy of the training paradigm, instead of the capacities of the models.
We propose to generate a symbolic and ordered sequence from the relation matrix
which is deterministic and easier for model to learn. Moreover, we design a
parallel row generation method to process overlong target sequences. Besides,
we introduce several negative sampling strategies to improve the performance
with balanced signals. Experimental results on four datasets show that our
proposed method can improve the performance of the generative DocRE models. We
have released our code at https://github.com/ayyyq/DORE.
Related papers
- COrAL: Order-Agnostic Language Modeling for Efficient Iterative Refinement [80.18490952057125]
Iterative refinement has emerged as an effective paradigm for enhancing the capabilities of large language models (LLMs) on complex tasks.
We propose Context-Wise Order-Agnostic Language Modeling (COrAL) to overcome these challenges.
Our approach models multiple token dependencies within manageable context windows, enabling the model to perform iterative refinement internally.
arXiv Detail & Related papers (2024-10-12T23:56:19Z) - Fine-Tuning with Divergent Chains of Thought Boosts Reasoning Through Self-Correction in Language Models [63.36637269634553]
We present a novel method of further improving performance by requiring models to compare multiple reasoning chains.
We find that instruction tuning on DCoT datasets boosts the performance of even smaller, and therefore more accessible, language models.
arXiv Detail & Related papers (2024-07-03T15:01:18Z) - How to Unleash the Power of Large Language Models for Few-shot Relation
Extraction? [28.413620806193165]
In this paper, we investigate principal methodologies, in-context learning and data generation, for few-shot relation extraction via GPT-3.5.
We observe that in-context learning can achieve performance on par with previous prompt learning approaches, and data generation with the large language model can boost previous solutions to obtain new state-of-the-art few-shot results.
arXiv Detail & Related papers (2023-05-02T15:55:41Z) - Towards Better Dynamic Graph Learning: New Architecture and Unified
Library [29.625205125350313]
DyGFormer is a Transformer-based architecture for dynamic graph learning.
DyGLib is a unified library with standard training pipelines and coding interfaces.
arXiv Detail & Related papers (2023-03-23T05:27:32Z) - AugTriever: Unsupervised Dense Retrieval and Domain Adaptation by Scalable Data Augmentation [44.93777271276723]
We propose two approaches that enable annotation-free and scalable training by creating pseudo querydocument pairs.
The query extraction method involves selecting salient spans from the original document to generate pseudo queries.
The transferred query generation method utilizes generation models trained for other NLP tasks, such as summarization, to produce pseudo queries.
arXiv Detail & Related papers (2022-12-17T10:43:25Z) - CodeExp: Explanatory Code Document Generation [94.43677536210465]
Existing code-to-text generation models produce only high-level summaries of code.
We conduct a human study to identify the criteria for high-quality explanatory docstring for code.
We present a multi-stage fine-tuning strategy and baseline models for the task.
arXiv Detail & Related papers (2022-11-25T18:05:44Z) - A sequence-to-sequence approach for document-level relation extraction [4.906513405712846]
Document-level relation extraction (DocRE) requires integrating information within and across sentences.
Seq2rel can learn the subtasks of DocRE end-to-end, replacing a pipeline of task-specific components.
arXiv Detail & Related papers (2022-04-03T16:03:19Z) - Improving Non-autoregressive Generation with Mixup Training [51.61038444990301]
We present a non-autoregressive generation model based on pre-trained transformer models.
We propose a simple and effective iterative training method called MIx Source and pseudo Target.
Our experiments on three generation benchmarks including question generation, summarization and paraphrase generation, show that the proposed framework achieves the new state-of-the-art results.
arXiv Detail & Related papers (2021-10-21T13:04:21Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.