Using LSTM and SARIMA Models to Forecast Cluster CPU Usage
- URL: http://arxiv.org/abs/2007.08092v1
- Date: Thu, 16 Jul 2020 03:29:13 GMT
- Title: Using LSTM and SARIMA Models to Forecast Cluster CPU Usage
- Authors: Langston Nashold, Rayan Krishnan
- Abstract summary: This work seeks to predict one resource, CPU usage, over both a short term and long term time scale.
We apply these models to Azure data resampled to 20 minutes per data point with the goal of predicting usage over the next hour for the short-term task and for the next three days for the long-term task.
The SARIMA model outperformed the LSTM for the long term prediction task, but performed poorer on the short term task.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As large scale cloud computing centers become more popular than individual
servers, predicting future resource demand need has become an important
problem. Forecasting resource need allows public cloud providers to proactively
allocate or deallocate resources for cloud services. This work seeks to predict
one resource, CPU usage, over both a short term and long term time scale.
To gain insight into the model characteristics that best support specific
tasks, we consider two vastly different architectures: the historically
relevant SARIMA model and the more modern neural network, LSTM model. We apply
these models to Azure data resampled to 20 minutes per data point with the goal
of predicting usage over the next hour for the short-term task and for the next
three days for the long-term task. The SARIMA model outperformed the LSTM for
the long term prediction task, but performed poorer on the short term task.
Furthermore, the LSTM model was more robust, whereas the SARIMA model relied on
the data meeting certain assumptions about seasonality.
Related papers
- Tackling Data Heterogeneity in Federated Time Series Forecasting [61.021413959988216]
Time series forecasting plays a critical role in various real-world applications, including energy consumption prediction, disease transmission monitoring, and weather forecasting.
Most existing methods rely on a centralized training paradigm, where large amounts of data are collected from distributed devices to a central cloud server.
We propose a novel framework, Fed-TREND, to address data heterogeneity by generating informative synthetic data as auxiliary knowledge carriers.
arXiv Detail & Related papers (2024-11-24T04:56:45Z) - Time-MoE: Billion-Scale Time Series Foundation Models with Mixture of Experts [25.503695417712997]
Time-MoE is a scalable and unified architecture designed to pre-train larger, more capable forecasting foundation models.
Time-MoE enhances computational efficiency by activating only a subset of networks for each prediction.
For the first time, we scaled a time series foundation model up to 2.4 billion parameters, achieving significantly improved forecasting precision.
arXiv Detail & Related papers (2024-09-24T12:42:18Z) - F-FOMAML: GNN-Enhanced Meta-Learning for Peak Period Demand Forecasting with Proxy Data [65.6499834212641]
We formulate the demand prediction as a meta-learning problem and develop the Feature-based First-Order Model-Agnostic Meta-Learning (F-FOMAML) algorithm.
By considering domain similarities through task-specific metadata, our model improved generalization, where the excess risk decreases as the number of training tasks increases.
Compared to existing state-of-the-art models, our method demonstrates a notable improvement in demand prediction accuracy, reducing the Mean Absolute Error by 26.24% on an internal vending machine dataset and by 1.04% on the publicly accessible JD.com dataset.
arXiv Detail & Related papers (2024-06-23T21:28:50Z) - Timer: Generative Pre-trained Transformers Are Large Time Series Models [83.03091523806668]
This paper aims at the early development of large time series models (LTSM)
During pre-training, we curate large-scale datasets with up to 1 billion time points.
To meet diverse application needs, we convert forecasting, imputation, and anomaly detection of time series into a unified generative task.
arXiv Detail & Related papers (2024-02-04T06:55:55Z) - Lag-Llama: Towards Foundation Models for Probabilistic Time Series
Forecasting [54.04430089029033]
We present Lag-Llama, a general-purpose foundation model for time series forecasting based on a decoder-only transformer architecture.
Lag-Llama is pretrained on a large corpus of diverse time series data from several domains, and demonstrates strong zero-shot generalization capabilities.
When fine-tuned on relatively small fractions of such previously unseen datasets, Lag-Llama achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-10-12T12:29:32Z) - Pushing the Limits of Pre-training for Time Series Forecasting in the
CloudOps Domain [54.67888148566323]
We introduce three large-scale time series forecasting datasets from the cloud operations domain.
We show it is a strong zero-shot baseline and benefits from further scaling, both in model and dataset size.
Accompanying these datasets and results is a suite of comprehensive benchmark results comparing classical and deep learning baselines to our pre-trained method.
arXiv Detail & Related papers (2023-10-08T08:09:51Z) - 1D-CapsNet-LSTM: A Deep Learning-Based Model for Multi-Step Stock Index
Forecasting [6.05458608266581]
This study investigates the potential of integrating a 1D CapsNet with an LSTM network for multi-step stock index forecasting.
To this end, a hybrid 1D-CapsNet-LSTM model is introduced, which utilizes a 1D CapsNet to generate high-level capsules.
The proposed 1D-CapsNet-LSTM model consistently outperforms baseline models in two key aspects.
arXiv Detail & Related papers (2023-10-03T14:33:34Z) - Online Evolutionary Neural Architecture Search for Multivariate
Non-Stationary Time Series Forecasting [72.89994745876086]
This work presents the Online Neuro-Evolution-based Neural Architecture Search (ONE-NAS) algorithm.
ONE-NAS is a novel neural architecture search method capable of automatically designing and dynamically training recurrent neural networks (RNNs) for online forecasting tasks.
Results demonstrate that ONE-NAS outperforms traditional statistical time series forecasting methods.
arXiv Detail & Related papers (2023-02-20T22:25:47Z) - HigeNet: A Highly Efficient Modeling for Long Sequence Time Series
Prediction in AIOps [30.963758935255075]
In this paper, we propose a highly efficient model named HigeNet to predict the long-time sequence time series.
We show that training time, resource usage and accuracy of the model are found to be significantly better than five state-of-the-art competing models.
arXiv Detail & Related papers (2022-11-13T13:48:43Z) - Pre-training Enhanced Spatial-temporal Graph Neural Network for
Multivariate Time Series Forecasting [13.441945545904504]
We propose a novel framework, in which STGNN is Enhanced by a scalable time series Pre-training model (STEP)
Specifically, we design a pre-training model to efficiently learn temporal patterns from very long-term history time series.
Our framework is capable of significantly enhancing downstream STGNNs, and our pre-training model aptly captures temporal patterns.
arXiv Detail & Related papers (2022-06-18T04:24:36Z) - Clustering Enabled Few-Shot Load Forecasting [2.0810096547938164]
We consider the load forecasting for a new user by observing only few shots (data points) of its energy consumption.
This task is challenging since the limited samples are insufficient to exploit the temporal characteristics.
We propose to utilize the historical load profile data from existing users to conduct effective clustering.
arXiv Detail & Related papers (2022-02-16T09:09:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.