Revisiting Transformer-based Models for Long Document Classification
- URL: http://arxiv.org/abs/2204.06683v1
- Date: Thu, 14 Apr 2022 00:44:36 GMT
- Title: Revisiting Transformer-based Models for Long Document Classification
- Authors: Xiang Dai and Ilias Chalkidis and Sune Darkner and Desmond Elliott
- Abstract summary: In real-world applications, multi-page multi-paragraph documents are common and cannot be efficiently encoded by vanilla Transformer-based models.
We compare different Transformer-based Long Document Classification (TrLDC) approaches that aim to mitigate the computational overhead of vanilla transformers.
We observe a clear benefit from being able to process longer text, and, based on our results, we derive practical advice of applying Transformer-based models on long document classification tasks.
- Score: 31.60414185940218
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The recent literature in text classification is biased towards short text
sequences (e.g., sentences or paragraphs). In real-world applications,
multi-page multi-paragraph documents are common and they cannot be efficiently
encoded by vanilla Transformer-based models. We compare different
Transformer-based Long Document Classification (TrLDC) approaches that aim to
mitigate the computational overhead of vanilla transformers to encode much
longer text, namely sparse attention and hierarchical encoding methods. We
examine several aspects of sparse attention (e.g., size of local attention
window, use of global attention) and hierarchical (e.g., document splitting
strategy) transformers on four document classification datasets covering
different domains. We observe a clear benefit from being able to process longer
text, and, based on our results, we derive practical advice of applying
Transformer-based models on long document classification tasks.
Related papers
- HDT: Hierarchical Document Transformer [70.2271469410557]
HDT exploits document structure by introducing auxiliary anchor tokens and redesigning the attention mechanism into a sparse multi-level hierarchy.
We develop a novel sparse attention kernel that considers the hierarchical structure of documents.
arXiv Detail & Related papers (2024-07-11T09:28:04Z) - Long-Range Transformer Architectures for Document Understanding [1.9331361036118608]
Document Understanding (DU) was not left behind with first Transformer based models for DU dating from late 2019.
We introduce 2 new multi-modal (text + layout) long-range models for DU based on efficient implementations of Transformers for long sequences.
Relative 2D attention revealed to be effective on dense text for both normal and long-range models.
arXiv Detail & Related papers (2023-09-11T14:45:24Z) - Can Model Fusing Help Transformers in Long Document Classification? An
Empirical Study [11.395215994671863]
Adapting NLP to multiple domains has introduced many new challenges for text classification.
The majority of the transformer models are limited to 512 tokens, and therefore, they struggle with long document classification problems.
In this research, we explore on employing Model Fusing for long document classification while comparing the results with well-known BERT and Longformer architectures.
arXiv Detail & Related papers (2023-07-18T18:21:26Z) - Hierarchical Local-Global Transformer for Temporal Sentence Grounding [58.247592985849124]
This paper studies the multimedia problem of temporal sentence grounding.
It aims to accurately determine the specific video segment in an untrimmed video according to a given sentence query.
arXiv Detail & Related papers (2022-08-31T14:16:56Z) - Long Document Summarization with Top-down and Bottom-up Inference [113.29319668246407]
We propose a principled inference framework to improve summarization models on two aspects.
Our framework assumes a hierarchical latent structure of a document where the top-level captures the long range dependency.
We demonstrate the effectiveness of the proposed framework on a diverse set of summarization datasets.
arXiv Detail & Related papers (2022-03-15T01:24:51Z) - Hierarchical Neural Network Approaches for Long Document Classification [3.6700088931938835]
We employ pre-trained Universal Sentence (USE) and Bidirectional Representations from Transformers (BERT) in a hierarchical setup to capture better representations efficiently.
Our proposed models are conceptually simple where we divide the input data into chunks and then pass this through base models of BERT and USE.
We show that USE + CNN/LSTM performs better than its stand-alone baseline. Whereas the BERT + CNN/LSTM performs on par with its stand-alone counterpart.
arXiv Detail & Related papers (2022-01-18T07:17:40Z) - HETFORMER: Heterogeneous Transformer with Sparse Attention for Long-Text
Extractive Summarization [57.798070356553936]
HETFORMER is a Transformer-based pre-trained model with multi-granularity sparse attentions for extractive summarization.
Experiments on both single- and multi-document summarization tasks show that HETFORMER achieves state-of-the-art performance in Rouge F1.
arXiv Detail & Related papers (2021-10-12T22:42:31Z) - Hi-Transformer: Hierarchical Interactive Transformer for Efficient and
Effective Long Document Modeling [51.79399904527525]
We propose a hierarchical interactive Transformer (Hi-Transformer) for efficient and effective long document modeling.
Hi-Transformer models documents in a hierarchical way, first learns sentence representations and then learns document representations.
Experiments on three benchmark datasets validate the efficiency and effectiveness of Hi-Transformer in long document modeling.
arXiv Detail & Related papers (2021-06-02T09:30:29Z) - Topical Change Detection in Documents via Embeddings of Long Sequences [4.13878392637062]
We formulate the task of text segmentation as an independent supervised prediction task.
By fine-tuning on paragraphs of similar sections, we are able to show that learned features encode topic information.
Unlike previous approaches, which mostly operate on sentence-level, we consistently use a broader context.
arXiv Detail & Related papers (2020-12-07T12:09:37Z) - Pretrained Transformers for Text Ranking: BERT and Beyond [53.83210899683987]
This survey provides an overview of text ranking with neural network architectures known as transformers.
The combination of transformers and self-supervised pretraining has been responsible for a paradigm shift in natural language processing.
arXiv Detail & Related papers (2020-10-13T15:20:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.