U-Mixer: An Unet-Mixer Architecture with Stationarity Correction for
Time Series Forecasting
- URL: http://arxiv.org/abs/2401.02236v1
- Date: Thu, 4 Jan 2024 12:41:40 GMT
- Title: U-Mixer: An Unet-Mixer Architecture with Stationarity Correction for
Time Series Forecasting
- Authors: Xiang Ma, Xuemei Li, Lexin Fang, Tianlong Zhao, Caiming Zhang
- Abstract summary: Non-stationarity in time series forecasting obstructs stable feature propagation through deep layers, disrupts feature distributions, and complicates learning data distribution changes.
We propose U-Mixer, which captures local temporal dependencies between different patches and channels separately.
We show that U-Mixer achieves 14.5% and 7.7% improvements over state-of-the-art (SOTA) methods.
- Score: 11.55346291812749
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Time series forecasting is a crucial task in various domains. Caused by
factors such as trends, seasonality, or irregular fluctuations, time series
often exhibits non-stationary. It obstructs stable feature propagation through
deep layers, disrupts feature distributions, and complicates learning data
distribution changes. As a result, many existing models struggle to capture the
underlying patterns, leading to degraded forecasting performance. In this
study, we tackle the challenge of non-stationarity in time series forecasting
with our proposed framework called U-Mixer. By combining Unet and Mixer,
U-Mixer effectively captures local temporal dependencies between different
patches and channels separately to avoid the influence of distribution
variations among channels, and merge low- and high-levels features to obtain
comprehensive data representations. The key contribution is a novel
stationarity correction method, explicitly restoring data distribution by
constraining the difference in stationarity between the data before and after
model processing to restore the non-stationarity information, while ensuring
the temporal dependencies are preserved. Through extensive experiments on
various real-world time series datasets, U-Mixer demonstrates its effectiveness
and robustness, and achieves 14.5\% and 7.7\% improvements over
state-of-the-art (SOTA) methods.
Related papers
- Learning Graph Structures and Uncertainty for Accurate and Calibrated Time-series Forecasting [65.40983982856056]
We introduce STOIC, that leverages correlations between time-series to learn underlying structure between time-series and to provide well-calibrated and accurate forecasts.
Over a wide-range of benchmark datasets STOIC provides 16% more accurate and better-calibrated forecasts.
arXiv Detail & Related papers (2024-07-02T20:14:32Z) - TimeMixer: Decomposable Multiscale Mixing for Time Series Forecasting [19.88184356154215]
Time series forecasting is widely used in applications, such as traffic planning and weather forecasting.
TimeMixer is able to achieve consistent state-of-the-art performances in both long-term and short-term forecasting tasks.
arXiv Detail & Related papers (2024-05-23T14:27:07Z) - CLeaRForecast: Contrastive Learning of High-Purity Representations for
Time Series Forecasting [2.5816901096123863]
Time series forecasting (TSF) holds significant importance in modern society, spanning numerous domains.
Previous representation learning-based TSF algorithms typically embrace a contrastive learning paradigm featuring segregated trend-periodicity representations.
We propose CLeaRForecast, a novel contrastive learning framework to learn high-purity time series representations with proposed sample, feature, and architecture purifying methods.
arXiv Detail & Related papers (2023-12-10T04:37:43Z) - One More Step: A Versatile Plug-and-Play Module for Rectifying Diffusion
Schedule Flaws and Enhancing Low-Frequency Controls [77.42510898755037]
One More Step (OMS) is a compact network that incorporates an additional simple yet effective step during inference.
OMS elevates image fidelity and harmonizes the dichotomy between training and inference, while preserving original model parameters.
Once trained, various pre-trained diffusion models with the same latent domain can share the same OMS module.
arXiv Detail & Related papers (2023-11-27T12:02:42Z) - Mitigating Shortcut Learning with Diffusion Counterfactuals and Diverse Ensembles [95.49699178874683]
We propose DiffDiv, an ensemble diversification framework exploiting Diffusion Probabilistic Models (DPMs)
We show that DPMs can generate images with novel feature combinations, even when trained on samples displaying correlated input features.
We show that DPM-guided diversification is sufficient to remove dependence on shortcut cues, without a need for additional supervised signals.
arXiv Detail & Related papers (2023-11-23T15:47:33Z) - A Multi-Scale Decomposition MLP-Mixer for Time Series Analysis [14.40202378972828]
We propose MSD-Mixer, a Multi-Scale Decomposition-Mixer, which learns to explicitly decompose and represent the input time series in its different layers.
We demonstrate that MSD-Mixer consistently and significantly outperforms other state-of-the-art algorithms with better efficiency.
arXiv Detail & Related papers (2023-10-18T13:39:07Z) - Generative Time Series Forecasting with Diffusion, Denoise, and
Disentanglement [51.55157852647306]
Time series forecasting has been a widely explored task of great importance in many applications.
It is common that real-world time series data are recorded in a short time period, which results in a big gap between the deep model and the limited and noisy time series.
We propose to address the time series forecasting problem with generative modeling and propose a bidirectional variational auto-encoder equipped with diffusion, denoise, and disentanglement.
arXiv Detail & Related papers (2023-01-08T12:20:46Z) - DRAformer: Differentially Reconstructed Attention Transformer for
Time-Series Forecasting [7.805077630467324]
Time-series forecasting plays an important role in many real-world scenarios, such as equipment life cycle forecasting, weather forecasting, and traffic flow forecasting.
It can be observed from recent research that a variety of transformer-based models have shown remarkable results in time-series forecasting.
However, there are still some issues that limit the ability of transformer-based models on time-series forecasting tasks.
arXiv Detail & Related papers (2022-06-11T10:34:29Z) - Multi-scale Attention Flow for Probabilistic Time Series Forecasting [68.20798558048678]
We propose a novel non-autoregressive deep learning model, called Multi-scale Attention Normalizing Flow(MANF)
Our model avoids the influence of cumulative error and does not increase the time complexity.
Our model achieves state-of-the-art performance on many popular multivariate datasets.
arXiv Detail & Related papers (2022-05-16T07:53:42Z) - TACTiS: Transformer-Attentional Copulas for Time Series [76.71406465526454]
estimation of time-varying quantities is a fundamental component of decision making in fields such as healthcare and finance.
We propose a versatile method that estimates joint distributions using an attention-based decoder.
We show that our model produces state-of-the-art predictions on several real-world datasets.
arXiv Detail & Related papers (2022-02-07T21:37:29Z) - Drift-Adjusted And Arbitrated Ensemble Framework For Time Series
Forecasting [0.491574468325115]
Time series forecasting is a challenging problem due to complex and evolving nature of time series data.
No one method is universally effective for all kinds of time series data.
We propose a re-weighting based method to adjust the assigned weights to various forecasters in order to account for such distribution-drift.
arXiv Detail & Related papers (2020-03-16T10:21:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.