Why does Prediction Accuracy Decrease over Time? Uncertain Positive
Learning for Cloud Failure Prediction
- URL: http://arxiv.org/abs/2402.00034v1
- Date: Mon, 8 Jan 2024 03:13:09 GMT
- Title: Why does Prediction Accuracy Decrease over Time? Uncertain Positive
Learning for Cloud Failure Prediction
- Authors: Haozhe Li, Minghua Ma, Yudong Liu, Pu Zhao, Lingling Zheng, Ze Li,
Yingnong Dang, Murali Chintalapati, Saravan Rajmohan, Qingwei Lin, Dongmei
Zhang
- Abstract summary: We find that the prediction accuracy may decrease by about 9% after retraining the models.
Considering that the mitigation actions may result in uncertain positive instances since they cannot be verified after mitigation, which may introduce more noise while updating the prediction model.
To tackle this problem, we design an Uncertain Positive Learning Risk Estimator (Uptake) approach.
- Score: 35.058991707881646
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the rapid growth of cloud computing, a variety of software services have
been deployed in the cloud. To ensure the reliability of cloud services, prior
studies focus on failure instance (disk, node, and switch, etc.) prediction.
Once the output of prediction is positive, mitigation actions are taken to
rapidly resolve the underlying failure. According to our real-world practice in
Microsoft Azure, we find that the prediction accuracy may decrease by about 9%
after retraining the models. Considering that the mitigation actions may result
in uncertain positive instances since they cannot be verified after mitigation,
which may introduce more noise while updating the prediction model. To the best
of our knowledge, we are the first to identify this Uncertain Positive Learning
(UPLearning) issue in the real-world cloud failure prediction scenario. To
tackle this problem, we design an Uncertain Positive Learning Risk Estimator
(Uptake) approach. Using two real-world datasets of disk failure prediction and
conducting node prediction experiments in Microsoft Azure, which is a top-tier
cloud provider that serves millions of users, we demonstrate Uptake can
significantly improve the failure prediction accuracy by 5% on average.
Related papers
- Next Best View For Point-Cloud Model Acquisition: Bayesian Approximation and Uncertainty Analysis [2.07180164747172]
This work adapts the point-net-based neural network for Next-Best-View (PC-NBV)
It incorporates dropout layers into the model's architecture, thus allowing the computation of the uncertainty estimate associated with its predictions.
The aim of the work is to improve the network's accuracy in correctly predicting the next best viewpoint.
arXiv Detail & Related papers (2024-11-04T01:32:09Z) - Building Defect Prediction Models by Online Learning Considering Defect Overlooking [1.5869998695491834]
Building defect prediction models based on online learning can enhance prediction accuracy.
A module predicted as "non-defective" can result in fewer test cases for such modules.
erroneous test results are used as learning data by online learning, which could negatively affect prediction accuracy.
arXiv Detail & Related papers (2024-04-17T03:20:46Z) - ExtremeCast: Boosting Extreme Value Prediction for Global Weather Forecast [57.6987191099507]
We introduce Exloss, a novel loss function that performs asymmetric optimization and highlights extreme values to obtain accurate extreme weather forecast.
We also introduce ExBooster, which captures the uncertainty in prediction outcomes by employing multiple random samples.
Our solution can achieve state-of-the-art performance in extreme weather prediction, while maintaining the overall forecast accuracy comparable to the top medium-range forecast models.
arXiv Detail & Related papers (2024-02-02T10:34:13Z) - Software Defect Prediction by Online Learning Considering Defect
Overlooking [1.655352281097533]
Building defect prediction models based on online learning can enhance prediction accuracy.
It continuously rebuilds a new prediction model when adding a new data point.
However, predicting a module as "non-defective" (i.e., negative prediction) can result in fewer test cases for such modules.
arXiv Detail & Related papers (2023-08-25T15:02:22Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - Learning to Predict Trustworthiness with Steep Slope Loss [69.40817968905495]
We study the problem of predicting trustworthiness on real-world large-scale datasets.
We observe that the trustworthiness predictors trained with prior-art loss functions are prone to view both correct predictions and incorrect predictions to be trustworthy.
We propose a novel steep slope loss to separate the features w.r.t. correct predictions from the ones w.r.t. incorrect predictions by two slide-like curves that oppose each other.
arXiv Detail & Related papers (2021-09-30T19:19:09Z) - Uncertainty-Aware Time-to-Event Prediction using Deep Kernel Accelerated
Failure Time Models [11.171712535005357]
We propose Deep Kernel Accelerated Failure Time models for the time-to-event prediction task.
Our model shows better point estimate performance than recurrent neural network based baselines in experiments on two real-world datasets.
arXiv Detail & Related papers (2021-07-26T14:55:02Z) - Learning Uncertainty with Artificial Neural Networks for Improved
Remaining Time Prediction of Business Processes [0.15229257192293202]
This paper is the first to apply these techniques to predictive process monitoring.
We found that they contribute towards more accurate predictions and work quickly.
This leads to many interesting applications, enables an earlier adoption of prediction systems with smaller datasets and fosters a better cooperation with humans.
arXiv Detail & Related papers (2021-05-12T10:18:57Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - Ambiguity in Sequential Data: Predicting Uncertain Futures with
Recurrent Models [110.82452096672182]
We propose an extension of the Multiple Hypothesis Prediction (MHP) model to handle ambiguous predictions with sequential data.
We also introduce a novel metric for ambiguous problems, which is better suited to account for uncertainties.
arXiv Detail & Related papers (2020-03-10T09:15:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.