Writing Style Aware Document-level Event Extraction
- URL: http://arxiv.org/abs/2201.03188v1
- Date: Mon, 10 Jan 2022 06:54:06 GMT
- Title: Writing Style Aware Document-level Event Extraction
- Authors: Zhuo Xu, Yue Wang, Lu Bai, Lixin Cui
- Abstract summary: Event extraction technology aims to automatically get the structural information from documents.
Most existing works discuss this issue by distinguishing the tokens as different roles while ignoring the writing styles of documents.
We argue that the writing style contains important clues for judging the roles for tokens and the ignorance of such patterns might lead to the performance degradation.
- Score: 11.146719375024674
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Event extraction, the technology that aims to automatically get the
structural information from documents, has attracted more and more attention in
many fields. Most existing works discuss this issue with the token-level
multi-label classification framework by distinguishing the tokens as different
roles while ignoring the writing styles of documents. The writing style is a
special way of content organizing for documents and it is relative fixed in
documents with a special field (e.g. financial, medical documents, etc.). We
argue that the writing style contains important clues for judging the roles for
tokens and the ignorance of such patterns might lead to the performance
degradation for the existing works. To this end, we model the writing style in
documents as a distribution of argument roles, i.e., Role-Rank Distribution,
and propose an event extraction model with the Role-Rank Distribution based
Supervision Mechanism to capture this pattern through the supervised training
process of an event extraction task. We compare our model with state-of-the-art
methods on several real-world datasets. The empirical results show that our
approach outperforms other alternatives with the captured patterns. This
verifies the writing style contains valuable information that could improve the
performance of the event extraction task.
Related papers
- Unified Multi-Modal Interleaved Document Representation for Information Retrieval [57.65409208879344]
We produce more comprehensive and nuanced document representations by holistically embedding documents interleaved with different modalities.
Specifically, we achieve this by leveraging the capability of recent vision-language models that enable the processing and integration of text, images, and tables into a unified format and representation.
arXiv Detail & Related papers (2024-10-03T17:49:09Z) - Leveraging Collection-Wide Similarities for Unsupervised Document Structure Extraction [61.998789448260005]
We propose to identify the typical structure of document within a collection.
We abstract over arbitrary header paraphrases, and ground each topic to respective document locations.
We develop an unsupervised graph-based method which leverages both inter- and intra-document similarities.
arXiv Detail & Related papers (2024-02-21T16:22:21Z) - On Task-personalized Multimodal Few-shot Learning for Visually-rich
Document Entity Retrieval [59.25292920967197]
Few-shot document entity retrieval (VDER) is an important topic in industrial NLP applications.
FewVEX is a new dataset to boost future research in the field of entity-level few-shot VDER.
We present a task-aware meta-learning based framework, with a central focus on achieving effective task personalization.
arXiv Detail & Related papers (2023-11-01T17:51:43Z) - Unified Pretraining Framework for Document Understanding [52.224359498792836]
We present UDoc, a new unified pretraining framework for document understanding.
UDoc is designed to support most document understanding tasks, extending the Transformer to take multimodal embeddings as input.
An important feature of UDoc is that it learns a generic representation by making use of three self-supervised losses.
arXiv Detail & Related papers (2022-04-22T21:47:04Z) - Synthetic Document Generator for Annotation-free Layout Recognition [15.657295650492948]
We describe a synthetic document generator that automatically produces realistic documents with labels for spatial positions, extents and categories of layout elements.
We empirically illustrate that a deep layout detection model trained purely on the synthetic documents can match the performance of a model that uses real documents.
arXiv Detail & Related papers (2021-11-11T01:58:44Z) - Event Extraction by Associating Event Types and Argument Roles [26.877240015683636]
Event extraction (EE) can be divided into two sub-tasks: event type classification and element extraction.
This paper proposes a novel neural association framework for the EE task.
Experimental results show that our approach consistently outperforms most state-of-the-art EE methods in both sub-tasks.
arXiv Detail & Related papers (2021-08-23T10:09:39Z) - SelfDoc: Self-Supervised Document Representation Learning [46.22910270334824]
SelfDoc is a task-agnostic pre-training framework for document image understanding.
Our framework exploits the positional, textual, and visual information of every semantically meaningful component in a document.
It achieves superior performance on multiple downstream tasks with significantly fewer document images used in the pre-training stage compared to previous works.
arXiv Detail & Related papers (2021-06-07T04:19:49Z) - DOC2PPT: Automatic Presentation Slides Generation from Scientific
Documents [76.19748112897177]
We present a novel task and approach for document-to-slide generation.
We propose a hierarchical sequence-to-sequence approach to tackle our task in an end-to-end manner.
Our approach exploits the inherent structures within documents and slides and incorporates paraphrasing and layout prediction modules to generate slides.
arXiv Detail & Related papers (2021-01-28T03:21:17Z) - Robust Document Representations using Latent Topics and Metadata [17.306088038339336]
We propose a novel approach to fine-tuning a pre-trained neural language model for document classification problems.
We generate document representations that capture both text and metadata artifacts in a task manner.
Our solution also incorporates metadata explicitly rather than just augmenting them with text.
arXiv Detail & Related papers (2020-10-23T21:52:38Z) - Towards a Multi-modal, Multi-task Learning based Pre-training Framework
for Document Representation Learning [5.109216329453963]
We introduce Document Topic Modelling and Document Shuffle Prediction as novel pre-training tasks.
We utilize the Longformer network architecture as the backbone to encode the multi-modal information from multi-page documents in an end-to-end fashion.
arXiv Detail & Related papers (2020-09-30T05:39:04Z) - Learning to Select Bi-Aspect Information for Document-Scale Text Content
Manipulation [50.01708049531156]
We focus on a new practical task, document-scale text content manipulation, which is the opposite of text style transfer.
In detail, the input is a set of structured records and a reference text for describing another recordset.
The output is a summary that accurately describes the partial content in the source recordset with the same writing style of the reference.
arXiv Detail & Related papers (2020-02-24T12:52:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.