Optimal starting point for time series forecasting
- URL: http://arxiv.org/abs/2409.16843v1
- Date: Wed, 25 Sep 2024 11:51:00 GMT
- Title: Optimal starting point for time series forecasting
- Authors: Yiming Zhong, Yinuo Ren, Guangyao Cao, Feng Li, Haobo Qi,
- Abstract summary: We introduce a novel approach called Optimal Starting Point Time Series Forecast (OSP-TSP)
By adjusting the sequence length via leveraging the XGBoost and LightGBM models, the proposed approach can determine optimal starting point (OSP) of the time series.
Empirical results indicate that predictions based on the OSP-TSP approach consistently outperform those using the complete dataset.
- Score: 1.9937737230710553
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent advances on time series forecasting mainly focus on improving the forecasting models themselves. However, managing the length of the input data can also significantly enhance prediction performance. In this paper, we introduce a novel approach called Optimal Starting Point Time Series Forecast (OSP-TSP) to capture the intrinsic characteristics of time series data. By adjusting the sequence length via leveraging the XGBoost and LightGBM models, the proposed approach can determine optimal starting point (OSP) of the time series and thus enhance the prediction performances. The performances of the OSP-TSP approach are then evaluated across various frequencies on the M4 dataset and other real-world datasets. Empirical results indicate that predictions based on the OSP-TSP approach consistently outperform those using the complete dataset. Moreover, recognizing the necessity of sufficient data to effectively train models for OSP identification, we further propose targeted solutions to address the issue of data insufficiency.
Related papers
- F-FOMAML: GNN-Enhanced Meta-Learning for Peak Period Demand Forecasting with Proxy Data [65.6499834212641]
We formulate the demand prediction as a meta-learning problem and develop the Feature-based First-Order Model-Agnostic Meta-Learning (F-FOMAML) algorithm.
By considering domain similarities through task-specific metadata, our model improved generalization, where the excess risk decreases as the number of training tasks increases.
Compared to existing state-of-the-art models, our method demonstrates a notable improvement in demand prediction accuracy, reducing the Mean Absolute Error by 26.24% on an internal vending machine dataset and by 1.04% on the publicly accessible JD.com dataset.
arXiv Detail & Related papers (2024-06-23T21:28:50Z) - Hyperparameter Tuning MLPs for Probabilistic Time Series Forecasting [6.579888565581481]
Time series forecasting attempts to predict future events by analyzing past trends and patterns.
Although well researched, certain critical aspects to the use of deep learning in time series forecasting remain ambiguous.
In this work, we introduce the largest metadataset for timeseries forecasting to date, named TSBench, comprising 97200 evaluations.
arXiv Detail & Related papers (2024-03-07T13:22:25Z) - Embedded feature selection in LSTM networks with multi-objective
evolutionary ensemble learning for time series forecasting [49.1574468325115]
We present a novel feature selection method embedded in Long Short-Term Memory networks.
Our approach optimize the weights and biases of the LSTM in a partitioned manner.
Experimental evaluations on air quality time series data from Italy and southeast Spain demonstrate that our method substantially improves the ability generalization of conventional LSTMs.
arXiv Detail & Related papers (2023-12-29T08:42:10Z) - Enhancing Algorithm Performance Understanding through tsMorph: Generating Semi-Synthetic Time Series for Robust Forecasting Evaluation [0.393259574660092]
We present tsMorph, a tool for generating semi-synthetic time series through dataset morphing.
In this paper, we show the benefits of tsMorph by assessing the predictive performance of the Long Short-Term Memory Network and DeepAR forecasting algorithms.
arXiv Detail & Related papers (2023-12-03T10:40:07Z) - Pushing the Limits of Pre-training for Time Series Forecasting in the
CloudOps Domain [54.67888148566323]
We introduce three large-scale time series forecasting datasets from the cloud operations domain.
We show it is a strong zero-shot baseline and benefits from further scaling, both in model and dataset size.
Accompanying these datasets and results is a suite of comprehensive benchmark results comparing classical and deep learning baselines to our pre-trained method.
arXiv Detail & Related papers (2023-10-08T08:09:51Z) - QBSD: Quartile-Based Seasonality Decomposition for Cost-Effective RAN KPI Forecasting [0.18416014644193066]
We introduce QBSD, a live single-step forecasting approach tailored to optimize the trade-off between accuracy and computational complexity.
QBSD has shown significant success with our real network RAN datasets of over several thousand cells.
Results demonstrate that the proposed method excels in runtime efficiency compared to the leading algorithms available.
arXiv Detail & Related papers (2023-06-09T15:59:27Z) - Exogenous Data in Forecasting: FARM -- A New Measure for Relevance
Evaluation [62.997667081978825]
We introduce a new approach named FARM - Forward Relevance Aligned Metric.
Our forward method relies on an angular measure that compares changes in subsequent data points to align time-warped series.
As a first validation step, we present the application of our FARM approach to synthetic but representative signals.
arXiv Detail & Related papers (2023-04-21T15:22:33Z) - Prediction-Oriented Bayesian Active Learning [51.426960808684655]
Expected predictive information gain (EPIG) is an acquisition function that measures information gain in the space of predictions rather than parameters.
EPIG leads to stronger predictive performance compared with BALD across a range of datasets and models.
arXiv Detail & Related papers (2023-04-17T10:59:57Z) - A Meta-Learning Approach to Predicting Performance and Data Requirements [163.4412093478316]
We propose an approach to estimate the number of samples required for a model to reach a target performance.
We find that the power law, the de facto principle to estimate model performance, leads to large error when using a small dataset.
We introduce a novel piecewise power law (PPL) that handles the two data differently.
arXiv Detail & Related papers (2023-03-02T21:48:22Z) - Hierarchical Proxy Modeling for Improved HPO in Time Series Forecasting [9.906423777470737]
We propose a novel technique, H-Pro, to drive HPO via test proxies by exploiting data hierarchies associated with time series datasets.
H-Pro can be applied on any off-the-shelf machine learning model to perform HPO.
Our approach outperforms existing state-of-the-art methods in Tourism, Wiki, and Traffic datasets.
arXiv Detail & Related papers (2022-11-28T06:37:15Z) - Financial Time Series Data Augmentation with Generative Adversarial
Networks and Extended Intertemporal Return Plots [2.365537081046599]
We apply state-of-the art image-based generative models for the task of data augmentation.
We introduce the extended intertemporal return plot (XIRP), a new image representation for time series.
Our approach proves to be effective in reducing the return forecast error by 7% on 79% of the financial data sets.
arXiv Detail & Related papers (2022-05-18T13:39:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.