tsMorph: generation of semi-synthetic time series to understand
algorithm performance
- URL: http://arxiv.org/abs/2312.01344v1
- Date: Sun, 3 Dec 2023 10:40:07 GMT
- Title: tsMorph: generation of semi-synthetic time series to understand
algorithm performance
- Authors: Mois\'es Santos and Andr\'e de Carvalho and Carlos Soares
- Abstract summary: We present tsMorph, a straightforward approach for generating semi-synthetic time series through dataset morphing.
In this paper, we demonstrate the utility of tsMorph by assessing the performance of the Long Short-Term Memory Network forecasting algorithm.
- Score: 0.4419843514606336
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Time series forecasting is a subject of significant scientific and industrial
importance. Despite the widespread utilization of forecasting methods, there is
a dearth of research aimed at comprehending the conditions under which these
methods yield favorable or unfavorable performances. Empirical studies,
although common, encounter challenges due to the limited availability of
datasets, impeding the extraction of reliable insights. To address this, we
present tsMorph, a straightforward approach for generating semi-synthetic time
series through dataset morphing. tsMorph operates by creating a sequence of
datasets derived from two original datasets. These newly generated datasets
exhibit a progressive departure from the characteristics of one dataset and a
convergence toward the attributes of the other. This method provides a valuable
alternative for obtaining substantial datasets. In this paper, we demonstrate
the utility of tsMorph by assessing the performance of the Long Short-Term
Memory Network forecasting algorithm. The time series under examination are
sourced from the NN5 Competition. The findings reveal compelling insights.
Notably, the performance of the Long Short-Term Memory Network improves
proportionally with the frequency of the time series. These experiments affirm
that tsMorph serves as an effective tool for gaining an understanding of
forecasting algorithm behaviors, offering a pathway to overcome the limitations
posed by empirical studies and enabling more extensive and reliable
experimentation.
Related papers
- State Sequences Prediction via Fourier Transform for Representation
Learning [111.82376793413746]
We propose State Sequences Prediction via Fourier Transform (SPF), a novel method for learning expressive representations efficiently.
We theoretically analyze the existence of structural information in state sequences, which is closely related to policy performance and signal regularity.
Experiments demonstrate that the proposed method outperforms several state-of-the-art algorithms in terms of both sample efficiency and performance.
arXiv Detail & Related papers (2023-10-24T14:47:02Z) - Pushing the Limits of Pre-training for Time Series Forecasting in the
CloudOps Domain [54.67888148566323]
We introduce three large-scale time series forecasting datasets from the cloud operations domain.
We show it is a strong zero-shot baseline and benefits from further scaling, both in model and dataset size.
Accompanying these datasets and results is a suite of comprehensive benchmark results comparing classical and deep learning baselines to our pre-trained method.
arXiv Detail & Related papers (2023-10-08T08:09:51Z) - MPR-Net:Multi-Scale Pattern Reproduction Guided Universality Time Series
Interpretable Forecasting [13.790498420659636]
Time series forecasting has received wide interest from existing research due to its broad applications inherent challenging.
This paper proposes a forecasting model, MPR-Net. It first adaptively decomposes multi-scale historical series patterns using convolution operation, then constructs a pattern extension forecasting method based on the prior knowledge of pattern reproduction, and finally reconstructs future patterns into future series using deconvolution operation.
By leveraging the temporal dependencies present in the time series, MPR-Net not only achieves linear time complexity, but also makes the forecasting process interpretable.
arXiv Detail & Related papers (2023-07-13T13:16:01Z) - Explainable Parallel RCNN with Novel Feature Representation for Time
Series Forecasting [0.0]
Time series forecasting is a fundamental challenge in data science.
We develop a parallel deep learning framework composed of RNN and CNN.
Extensive experiments on three datasets reveal the effectiveness of our method.
arXiv Detail & Related papers (2023-05-08T17:20:13Z) - Can LMs Generalize to Future Data? An Empirical Analysis on Text
Summarization [50.20034493626049]
Recent pre-trained language models (PLMs) achieve promising results in existing abstractive summarization datasets.
Existing summarization benchmarks overlap in time with the standard pre-training corpora and finetuning datasets.
We show that parametric knowledge stored in summarization models significantly affects the faithfulness of the generated summaries on future data.
arXiv Detail & Related papers (2023-05-03T08:08:07Z) - A Novel Method Combines Moving Fronts, Data Decomposition and Deep
Learning to Forecast Intricate Time Series [0.0]
Indian Summer Monsoon Rainfall (ISMR) is a very complex time series.
Conventional one-time decomposition technique suffers from a leak of information from the future.
Moving Front (MF) method is proposed to prevent data leakage.
arXiv Detail & Related papers (2023-03-11T12:07:26Z) - Convolutional generative adversarial imputation networks for
spatio-temporal missing data in storm surge simulations [86.5302150777089]
Generative Adversarial Imputation Nets (GANs) and GAN-based techniques have attracted attention as unsupervised machine learning methods.
We name our proposed method as Con Conval Generative Adversarial Imputation Nets (Conv-GAIN)
arXiv Detail & Related papers (2021-11-03T03:50:48Z) - Time Series is a Special Sequence: Forecasting with Sample Convolution
and Interaction [9.449017120452675]
Time series is a special type of sequence data, a set of observations collected at even intervals of time and ordered chronologically.
Existing deep learning techniques use generic sequence models for time series analysis, which ignore some of its unique properties.
We propose a novel neural network architecture and apply it for the time series forecasting problem, wherein we conduct sample convolution and interaction at multiple resolutions for temporal modeling.
arXiv Detail & Related papers (2021-06-17T08:15:04Z) - Towards Synthetic Multivariate Time Series Generation for Flare
Forecasting [5.098461305284216]
One of the limiting factors in training data-driven, rare-event prediction algorithms is the scarcity of the events of interest.
In this study, we explore the usefulness of the conditional generative adversarial network (CGAN) as a means to perform data-informed oversampling.
arXiv Detail & Related papers (2021-05-16T22:23:23Z) - Learning summary features of time series for likelihood free inference [93.08098361687722]
We present a data-driven strategy for automatically learning summary features from time series data.
Our results indicate that learning summary features from data can compete and even outperform LFI methods based on hand-crafted values.
arXiv Detail & Related papers (2020-12-04T19:21:37Z) - Convolutional Tensor-Train LSTM for Spatio-temporal Learning [116.24172387469994]
We propose a higher-order LSTM model that can efficiently learn long-term correlations in the video sequence.
This is accomplished through a novel tensor train module that performs prediction by combining convolutional features across time.
Our results achieve state-of-the-art performance-art in a wide range of applications and datasets.
arXiv Detail & Related papers (2020-02-21T05:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.