URegM: a unified prediction model of resource consumption for
refactoring software smells in open source cloud
- URL: http://arxiv.org/abs/2310.14444v1
- Date: Sun, 22 Oct 2023 23:03:35 GMT
- Title: URegM: a unified prediction model of resource consumption for
refactoring software smells in open source cloud
- Authors: Asif Imran and Tevfik Kosar
- Abstract summary: We propose a framework called Unified Regression Modelling (URegM) which predicts the impact of code smell on cloud resource usage.
Results show that URegM is capable of accurately predicting resource consumption due to code smell.
- Score: 3.9704849108478704
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The low cost and rapid provisioning capabilities have made the cloud a
desirable platform to launch complex scientific applications. However, resource
utilization optimization is a significant challenge for cloud service
providers, since the earlier focus is provided on optimizing resources for the
applications that run on the cloud, with a low emphasis being provided on
optimizing resource utilization of the cloud computing internal processes. Code
refactoring has been associated with improving the maintenance and
understanding of software code. However, analyzing the impact of the
refactoring source code of the cloud and studying its impact on cloud resource
usage require further analysis. In this paper, we propose a framework called
Unified Regression Modelling (URegM) which predicts the impact of code smell
refactoring on cloud resource usage. We test our experiments in a real-life
cloud environment using a complex scientific application as a workload. Results
show that URegM is capable of accurately predicting resource consumption due to
code smell refactoring. This will permit cloud service providers with advanced
knowledge about the impact of refactoring code smells on resource consumption,
thus allowing them to plan their resource provisioning and code refactoring
more effectively.
Related papers
- Reinforcement Learning-Based Adaptive Load Balancing for Dynamic Cloud Environments [0.0]
We propose a novel adaptive load balancing framework using Reinforcement Learning (RL) to address these challenges.
Our framework is designed to dynamically reallocate tasks to minimize latency and ensure balanced resource usage across servers.
Experimental results show that the proposed RL-based load balancer outperforms traditional algorithms in terms of response time, resource utilization, and adaptability to changing workloads.
arXiv Detail & Related papers (2024-09-07T19:40:48Z) - ReGAL: Refactoring Programs to Discover Generalizable Abstractions [59.05769810380928]
Generalizable Abstraction Learning (ReGAL) is a method for learning a library of reusable functions via codeization.
We find that the shared function libraries discovered by ReGAL make programs easier to predict across diverse domains.
For CodeLlama-13B, ReGAL results in absolute accuracy increases of 11.5% on LOGO, 26.1% on date understanding, and 8.1% on TextCraft, outperforming GPT-3.5 in two of three domains.
arXiv Detail & Related papers (2024-01-29T18:45:30Z) - LLM-Assisted Code Cleaning For Training Accurate Code Generators [53.087019724256606]
We investigate data quality for code and find that making the code more structured and readable leads to improved code generation performance of the system.
We build a novel data-cleaning pipeline that uses these principles to transform existing programs.
We evaluate our approach on two challenging algorithmic code generation benchmarks and find that fine-tuning CodeLLaMa-7B improves the performance by up to 30% compared to fine-tuning on the original dataset.
arXiv Detail & Related papers (2023-11-25T02:45:50Z) - CloudScent: a model for code smell analysis in open-source cloud [2.990411348977783]
We propose a model called CloudScent which is an open source mechanism to detect smells in open-source cloud.
Results show that CloudScent is capable of accurately detecting 8 code smells in cloud.
arXiv Detail & Related papers (2023-07-22T19:01:38Z) - Predicting the Impact of Batch Refactoring Code Smells on Application
Resource Consumption [3.5557219875516646]
This paper determines the relationship between software code smell batch, and resource consumption.
Next, it aims to design algorithms to predict the impact of code smell on resource consumption.
arXiv Detail & Related papers (2023-06-27T19:28:05Z) - Mining for Cost Awareness in the Infrastructure as Code Artifacts of Cloud-based Applications: an Exploratory Study [1.5612101323427952]
This study aims to provide empirical evidence of cost awareness by mining open source repositories of cloud-based applications.
A systematic search through 152,735 repositories resulted in the selection of 2,010 relevant ones.
The findings indicate that developers are not only concerned with the cost of their application deployments but also take actions to reduce these costs.
arXiv Detail & Related papers (2023-04-15T10:49:35Z) - Outsourcing Training without Uploading Data via Efficient Collaborative
Open-Source Sampling [49.87637449243698]
Traditional outsourcing requires uploading device data to the cloud server.
We propose to leverage widely available open-source data, which is a massive dataset collected from public and heterogeneous sources.
We develop a novel strategy called Efficient Collaborative Open-source Sampling (ECOS) to construct a proximal proxy dataset from open-source data for cloud training.
arXiv Detail & Related papers (2022-10-23T00:12:18Z) - Device-Cloud Collaborative Recommendation via Meta Controller [65.97416287295152]
We propose a meta controller to dynamically manage the collaboration between the on-device recommender and the cloud-based recommender.
On the basis of the counterfactual samples and the extended training, extensive experiments in the industrial recommendation scenarios show the promise of meta controller.
arXiv Detail & Related papers (2022-07-07T03:23:04Z) - Auto-Split: A General Framework of Collaborative Edge-Cloud AI [49.750972428032355]
This paper describes the techniques and engineering practice behind Auto-Split, an edge-cloud collaborative prototype of Huawei Cloud.
To the best of our knowledge, there is no existing industry product that provides the capability of Deep Neural Network (DNN) splitting.
arXiv Detail & Related papers (2021-08-30T08:03:29Z) - Machine Learning (ML)-Centric Resource Management in Cloud Computing: A
Review and Future Directions [22.779373079539713]
Infrastructure as a Service (I) is one of the most important and rapidly growing fields.
One of the most important aspects of cloud computing for I is resource management.
Machine learning is being used to handle a variety of resource management tasks.
arXiv Detail & Related papers (2021-05-09T08:03:58Z) - Pollux: Co-adaptive Cluster Scheduling for Goodput-Optimized Deep
Learning [61.29990368322931]
Pollux improves scheduling performance in deep learning (DL) clusters by adaptively co-optimizing inter-dependent factors.
Pollux reduces average job completion times by 37-50% relative to state-of-the-art DL schedulers.
arXiv Detail & Related papers (2020-08-27T16:56:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.