FECAM: Frequency Enhanced Channel Attention Mechanism for Time Series
Forecasting
- URL: http://arxiv.org/abs/2212.01209v1
- Date: Fri, 2 Dec 2022 14:40:55 GMT
- Title: FECAM: Frequency Enhanced Channel Attention Mechanism for Time Series
Forecasting
- Authors: Maowei Jiang, Pengyu Zeng, Kai Wang, Huan Liu, Wenbo Chen, Haoran Liu
- Abstract summary: Time series forecasting is a long-standing challenge due to the real-world information is in various scenario.
We believe it's the reason that model's lacking ability of capturing frequency information which richly contains in real world datasets.
We propose a novel frequency enhanced channel attention that adaptively modelling frequency interdependencies between channels based on Discrete Cosine Transform.
- Score: 21.933798421967225
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Time series forecasting is a long-standing challenge due to the real-world
information is in various scenario (e.g., energy, weather, traffic, economics,
earthquake warning). However some mainstream forecasting model forecasting
result is derailed dramatically from ground truth. We believe it's the reason
that model's lacking ability of capturing frequency information which richly
contains in real world datasets. At present, the mainstream frequency
information extraction methods are Fourier transform(FT) based. However, use of
FT is problematic due to Gibbs phenomenon. If the values on both sides of
sequences differ significantly, oscillatory approximations are observed around
both sides and high frequency noise will be introduced. Therefore We propose a
novel frequency enhanced channel attention that adaptively modelling frequency
interdependencies between channels based on Discrete Cosine Transform which
would intrinsically avoid high frequency noise caused by problematic periodity
during Fourier Transform, which is defined as Gibbs Phenomenon. We show that
this network generalize extremely effectively across six real-world datasets
and achieve state-of-the-art performance, we further demonstrate that frequency
enhanced channel attention mechanism module can be flexibly applied to
different networks. This module can improve the prediction ability of existing
mainstream networks, which reduces 35.99% MSE on LSTM, 10.01% on Reformer,
8.71% on Informer, 8.29% on Autoformer, 8.06% on Transformer, etc., at a slight
computational cost ,with just a few line of code. Our codes and data are
available at https://github.com/Zero-coder/FECAM.
Related papers
- Less is more: Embracing sparsity and interpolation with Esiformer for time series forecasting [19.8447763392479]
Time series data generated from real-world applications always exhibits high variance and lots of noise.
We propose the Esiformer, which apply on the original data, decreasing the overall variance of the data and alleviating the influence of noise.
Our method outperforms leading model PatchTST, reducing MSE by 6.5% and MAE by 5.8%.
arXiv Detail & Related papers (2024-10-08T06:45:47Z) - Accelerating Inference of Networks in the Frequency Domain [8.125023712173686]
We propose performing network inference in the frequency domain to speed up networks whose frequency parameters are sparse.
In particular, we propose a frequency inference chain that is dual to the network inference in the spatial domain.
The proposed approach significantly improves accuracy in the case of a high speedup ratio (over 100x)
arXiv Detail & Related papers (2024-10-06T03:34:38Z) - Deep Frequency Derivative Learning for Non-stationary Time Series Forecasting [12.989064148254936]
We present a deep frequency derivative learning framework, DERITS, for non-stationary time series forecasting.
Specifically, DERITS is built upon a novel reversible transformation, namely Frequency Derivative Transformation (FDT)
arXiv Detail & Related papers (2024-06-29T17:56:59Z) - Fredformer: Frequency Debiased Transformer for Time Series Forecasting [8.356290446630373]
The Transformer model has shown leading performance in time series forecasting.
It tends to learn low-frequency features in the data and overlook high-frequency features, showing a frequency bias.
We propose Fredformer, a framework designed to mitigate frequency bias by learning features equally across different frequency bands.
arXiv Detail & Related papers (2024-06-13T11:29:21Z) - Frequency-Aware Deepfake Detection: Improving Generalizability through
Frequency Space Learning [81.98675881423131]
This research addresses the challenge of developing a universal deepfake detector that can effectively identify unseen deepfake images.
Existing frequency-based paradigms have relied on frequency-level artifacts introduced during the up-sampling in GAN pipelines to detect forgeries.
We introduce a novel frequency-aware approach called FreqNet, centered around frequency domain learning, specifically designed to enhance the generalizability of deepfake detectors.
arXiv Detail & Related papers (2024-03-12T01:28:00Z) - WFTNet: Exploiting Global and Local Periodicity in Long-term Time Series
Forecasting [61.64303388738395]
We propose a Wavelet-Fourier Transform Network (WFTNet) for long-term time series forecasting.
Tests on various time series datasets show WFTNet consistently outperforms other state-of-the-art baselines.
arXiv Detail & Related papers (2023-09-20T13:44:18Z) - CARD: Channel Aligned Robust Blend Transformer for Time Series
Forecasting [50.23240107430597]
We design a special Transformer, i.e., Channel Aligned Robust Blend Transformer (CARD for short), that addresses key shortcomings of CI type Transformer in time series forecasting.
First, CARD introduces a channel-aligned attention structure that allows it to capture both temporal correlations among signals.
Second, in order to efficiently utilize the multi-scale knowledge, we design a token blend module to generate tokens with different resolutions.
Third, we introduce a robust loss function for time series forecasting to alleviate the potential overfitting issue.
arXiv Detail & Related papers (2023-05-20T05:16:31Z) - Transform Once: Efficient Operator Learning in Frequency Domain [69.74509540521397]
We study deep neural networks designed to harness the structure in frequency domain for efficient learning of long-range correlations in space or time.
This work introduces a blueprint for frequency domain learning through a single transform: transform once (T1)
arXiv Detail & Related papers (2022-11-26T01:56:05Z) - Global Filter Networks for Image Classification [90.81352483076323]
We present a conceptually simple yet computationally efficient architecture that learns long-term spatial dependencies in the frequency domain with log-linear complexity.
Our results demonstrate that GFNet can be a very competitive alternative to transformer-style models and CNNs in efficiency, generalization ability and robustness.
arXiv Detail & Related papers (2021-07-01T17:58:16Z) - Over-the-Air Federated Learning from Heterogeneous Data [107.05618009955094]
Federated learning (FL) is a framework for distributed learning of centralized models.
We develop a Convergent OTA FL (COTAF) algorithm which enhances the common local gradient descent (SGD) FL algorithm.
We numerically show that the precoding induced by COTAF notably improves the convergence rate and the accuracy of models trained via OTA FL.
arXiv Detail & Related papers (2020-09-27T08:28:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.