Forecasting Workload in Cloud Computing: Towards Uncertainty-Aware
Predictions and Transfer Learning
- URL: http://arxiv.org/abs/2303.13525v2
- Date: Sun, 12 Nov 2023 13:07:24 GMT
- Title: Forecasting Workload in Cloud Computing: Towards Uncertainty-Aware
Predictions and Transfer Learning
- Authors: Andrea Rossi and Andrea Visentin and Diego Carraro and Steven
Prestwich and Kenneth N. Brown
- Abstract summary: We show that modelling the uncertainty of predictions has a positive impact on performance.
We investigate whether our models benefit transfer learning capabilities across different domains.
- Score: 1.5749416770494704
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Predicting future resource demand in Cloud Computing is essential for
optimizing the trade-off between serving customers' requests efficiently and
minimizing the provisioning cost. Modelling prediction uncertainty is also
desirable to better inform the resource decision-making process, but research
in this field is under-investigated. In this paper, we propose univariate and
bivariate Bayesian deep learning models that provide predictions of future
workload demand and its uncertainty. We run extensive experiments on Google and
Alibaba clusters, where we first train our models with datasets from different
cloud providers and compare them with LSTM-based baselines. Results show that
modelling the uncertainty of predictions has a positive impact on performance,
especially on service level metrics, because uncertainty quantification can be
tailored to desired target service levels that are critical in cloud
applications. Moreover, we investigate whether our models benefit transfer
learning capabilities across different domains, i.e. dataset distributions.
Experiments on the same workload datasets reveal that acceptable transfer
learning performance can be achieved within the same provider (because
distributions are more similar). Also, domain knowledge does not transfer when
the source and target domains are very different (e.g. from different
providers), but this performance degradation can be mitigated by increasing the
training set size of the source domain.
Related papers
- Can Performant LLMs Be Ethical? Quantifying the Impact of Web Crawling Opt-Outs [42.58914814153536]
We quantify the performance difference between models trained on datasets that comply with web crawling opt-outs, and those that do not.
Our experiments with 1.5B models show that, as of January 2025, compliance with web data opt-outs does not degrade general knowledge acquisition.
However, in specialized domains such as biomedical research, excluding major publishers leads to performance declines.
arXiv Detail & Related papers (2025-04-08T17:08:06Z) - A Conformal Approach to Feature-based Newsvendor under Model Misspecification [2.801095519296785]
We propose a model-free and distribution-free framework inspired by conformal prediction.
We validate our framework using both simulated data and a real-world dataset from the Capital Bikeshare program in Washington, D.C.
arXiv Detail & Related papers (2024-12-17T18:34:43Z) - F-FOMAML: GNN-Enhanced Meta-Learning for Peak Period Demand Forecasting with Proxy Data [65.6499834212641]
We formulate the demand prediction as a meta-learning problem and develop the Feature-based First-Order Model-Agnostic Meta-Learning (F-FOMAML) algorithm.
By considering domain similarities through task-specific metadata, our model improved generalization, where the excess risk decreases as the number of training tasks increases.
Compared to existing state-of-the-art models, our method demonstrates a notable improvement in demand prediction accuracy, reducing the Mean Absolute Error by 26.24% on an internal vending machine dataset and by 1.04% on the publicly accessible JD.com dataset.
arXiv Detail & Related papers (2024-06-23T21:28:50Z) - Singular Value Penalization and Semantic Data Augmentation for Fully
Test-Time Adaptation [5.891527229524256]
Test-time adaptation (FTTA) adapts a model that is trained on a source domain to a target domain during the testing phase.
We propose maximizing the sum of singular values while minimizing their variance.
This enables the model's focus toward the smaller singular values, enhancing discriminability between more challenging classes and effectively increasing the diversity of prediction results.
arXiv Detail & Related papers (2023-12-10T01:08:56Z) - One-Shot Federated Learning with Classifier-Guided Diffusion Models [44.604485649167216]
One-shot federated learning (OSFL) has gained attention in recent years due to its low communication cost.
In this paper, we explore the novel opportunities that diffusion models bring to OSFL and propose FedCADO.
FedCADO generates data that complies with clients' distributions and subsequently training the aggregated model on the server.
arXiv Detail & Related papers (2023-11-15T11:11:25Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - Task-Aware Machine Unlearning and Its Application in Load Forecasting [4.00606516946677]
This paper introduces the concept of machine unlearning which is specifically designed to remove the influence of part of the dataset on an already trained forecaster.
A performance-aware algorithm is proposed by evaluating the sensitivity of local model parameter change using influence function and sample re-weighting.
We tested the unlearning algorithms on linear, CNN, andMixer based load forecasters with a realistic load dataset.
arXiv Detail & Related papers (2023-08-28T08:50:12Z) - On the Trade-off of Intra-/Inter-class Diversity for Supervised
Pre-training [72.8087629914444]
We study the impact of the trade-off between the intra-class diversity (the number of samples per class) and the inter-class diversity (the number of classes) of a supervised pre-training dataset.
With the size of the pre-training dataset fixed, the best downstream performance comes with a balance on the intra-/inter-class diversity.
arXiv Detail & Related papers (2023-05-20T16:23:50Z) - Data-Driven Offline Decision-Making via Invariant Representation
Learning [97.49309949598505]
offline data-driven decision-making involves synthesizing optimized decisions with no active interaction.
A key challenge is distributional shift: when we optimize with respect to the input into a model trained from offline data, it is easy to produce an out-of-distribution (OOD) input that appears erroneously good.
In this paper, we formulate offline data-driven decision-making as domain adaptation, where the goal is to make accurate predictions for the value of optimized decisions.
arXiv Detail & Related papers (2022-11-21T11:01:37Z) - Uncertainty-guided Source-free Domain Adaptation [77.3844160723014]
Source-free domain adaptation (SFDA) aims to adapt a classifier to an unlabelled target data set by only using a pre-trained source model.
We propose quantifying the uncertainty in the source model predictions and utilizing it to guide the target adaptation.
arXiv Detail & Related papers (2022-08-16T08:03:30Z) - Trust-Based Cloud Machine Learning Model Selection For Industrial IoT
and Smart City Services [5.333802479607541]
We consider the paradigm where cloud service providers collect big data from resource-constrained devices for building Machine Learning prediction models.
Our proposed solution comprises an intelligent-time reconfiguration that maximizes the level of trust of ML models.
Our results show that the selected model's trust level is 0.7% to 2.53% less compared to the results obtained using ILP.
arXiv Detail & Related papers (2020-08-11T23:58:03Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.