PETformer: Long-term Time Series Forecasting via Placeholder-enhanced
Transformer
- URL: http://arxiv.org/abs/2308.04791v2
- Date: Sat, 7 Oct 2023 09:24:40 GMT
- Title: PETformer: Long-term Time Series Forecasting via Placeholder-enhanced
Transformer
- Authors: Shengsheng Lin, Weiwei Lin, Wentai Wu, Songbo Wang, Yongxiang Wang
- Abstract summary: This study investigates key issues when applying Transformer to long-term time series forecasting tasks.
We introduce the Placeholder-enhanced Technique (PET) to enhance the computational efficiency and predictive accuracy of Transformer in LTSF tasks.
PETformer achieves state-of-the-art performance on eight commonly used public datasets for LTSF, surpassing all existing models.
- Score: 5.095882718779794
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, the superiority of Transformer for long-term time series
forecasting (LTSF) tasks has been challenged, particularly since recent work
has shown that simple models can outperform numerous Transformer-based
approaches. This suggests that a notable gap remains in fully leveraging the
potential of Transformer in LTSF tasks. Consequently, this study investigates
key issues when applying Transformer to LTSF, encompassing aspects of temporal
continuity, information density, and multi-channel relationships. We introduce
the Placeholder-enhanced Technique (PET) to enhance the computational
efficiency and predictive accuracy of Transformer in LTSF tasks. Furthermore,
we delve into the impact of larger patch strategies and channel interaction
strategies on Transformer's performance, specifically Long Sub-sequence
Division (LSD) and Multi-channel Separation and Interaction (MSI). These
strategies collectively constitute a novel model termed PETformer. Extensive
experiments have demonstrated that PETformer achieves state-of-the-art
performance on eight commonly used public datasets for LTSF, surpassing all
existing models. The insights and enhancement methodologies presented in this
paper serve as valuable reference points and sources of inspiration for future
research endeavors.
Related papers
- TEAFormers: TEnsor-Augmented Transformers for Multi-Dimensional Time Series Forecasting [14.43696537295348]
Multi-dimensional time series data are increasingly prevalent in fields such as economics, finance, and climate science.
Traditional Transformer models, though adept with sequential data, do not effectively preserve these multi-dimensional structures.
We introduce the vectors-Augmented Transformer (TEAFormer), a novel method that incorporates tensor expansion and compression within the Transformer framework.
arXiv Detail & Related papers (2024-10-27T13:32:12Z) - PDMLP: Patch-based Decomposed MLP for Long-Term Time Series Forecasting [0.0]
Recent studies have attempted to refine the Transformer architecture to demonstrate its effectiveness in Long-Term Time Series Forecasting (LTSF) tasks.
We attribute the effectiveness of these models largely to the adopted Patch mechanism, which enhances sequence locality.
Further investigation suggests that simple linear layers augmented with the Patch mechanism may outperform complex Transformer-based LTSF models.
arXiv Detail & Related papers (2024-05-22T12:12:20Z) - A Systematic Review for Transformer-based Long-term Series Forecasting [7.414422194379818]
Transformer architecture has proven to be the most successful solution to extract semantic correlations.
Various variants have enabled transformer architecture to handle long-term time series forecasting tasks.
arXiv Detail & Related papers (2023-10-31T06:37:51Z) - U-shaped Transformer: Retain High Frequency Context in Time Series
Analysis [0.5710971447109949]
In this paper, we consider the low-pass characteristics of transformers and try to incorporate the advantages of them.
We introduce patch merge and split operation to extract features with different scales and use larger datasets to fully make use of the transformer backbone.
Our experiments demonstrate that the model performs at an advanced level across multiple datasets with relatively low cost.
arXiv Detail & Related papers (2023-07-18T07:15:26Z) - Emergent Agentic Transformer from Chain of Hindsight Experience [96.56164427726203]
We show that a simple transformer-based model performs competitively with both temporal-difference and imitation-learning-based approaches.
This is the first time that a simple transformer-based model performs competitively with both temporal-difference and imitation-learning-based approaches.
arXiv Detail & Related papers (2023-05-26T00:43:02Z) - Towards Long-Term Time-Series Forecasting: Feature, Pattern, and
Distribution [57.71199089609161]
Long-term time-series forecasting (LTTF) has become a pressing demand in many applications, such as wind power supply planning.
Transformer models have been adopted to deliver high prediction capacity because of the high computational self-attention mechanism.
We propose an efficient Transformerbased model, named Conformer, which differentiates itself from existing methods for LTTF in three aspects.
arXiv Detail & Related papers (2023-01-05T13:59:29Z) - A K-variate Time Series Is Worth K Words: Evolution of the Vanilla
Transformer Architecture for Long-term Multivariate Time Series Forecasting [52.33042819442005]
Transformer has become the de facto solution for MTSF, especially for the long-term cases.
In this study, we point out that the current tokenization strategy in MTSF Transformer architectures ignores the token inductive bias of Transformers.
We make a series of evolution on the basic architecture of the vanilla MTSF transformer.
Surprisingly, the evolved simple transformer architecture is highly effective, which successfully avoids the over-smoothing phenomena in the vanilla MTSF transformer.
arXiv Detail & Related papers (2022-12-06T07:00:31Z) - CLMFormer: Mitigating Data Redundancy to Revitalize Transformer-based
Long-Term Time Series Forecasting System [46.39662315849883]
Long-term time-series forecasting (LTSF) plays a crucial role in various practical applications.
Existing Transformer-based models, such as Fedformer and Informer, often achieve their best performances on validation sets after just a few epochs.
We propose a novel approach to address this issue by employing curriculum learning and introducing a memory-driven decoder.
arXiv Detail & Related papers (2022-07-16T04:05:15Z) - Augmented Shortcuts for Vision Transformers [49.70151144700589]
We study the relationship between shortcuts and feature diversity in vision transformer models.
We present an augmented shortcut scheme, which inserts additional paths with learnable parameters in parallel on the original shortcuts.
Experiments conducted on benchmark datasets demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2021-06-30T09:48:30Z) - Applying the Transformer to Character-level Transduction [68.91664610425114]
The transformer has been shown to outperform recurrent neural network-based sequence-to-sequence models in various word-level NLP tasks.
We show that with a large enough batch size, the transformer does indeed outperform recurrent models for character-level tasks.
arXiv Detail & Related papers (2020-05-20T17:25:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.