Combination of Convolutional Neural Network and Gated Recurrent Unit for
Energy Aware Resource Allocation
- URL: http://arxiv.org/abs/2106.12178v1
- Date: Wed, 23 Jun 2021 05:57:51 GMT
- Title: Combination of Convolutional Neural Network and Gated Recurrent Unit for
Energy Aware Resource Allocation
- Authors: Zeinab Khodaverdian, Hossein Sadr, Seyed Ahmad Edalatpanah and Mojdeh
Nazari Solimandarabi
- Abstract summary: Cloud computing service models have experienced rapid growth and inefficient resource usage is one of the greatest causes of high energy consumption in cloud data centers.
Resource allocation in cloud data centers aiming to reduce energy consumption has been conducted using live migration of Virtual Machines (VMs) and their consolidation into the small number of Physical Machines (PMs)
To solve this issue, can be classified according to the pattern of user requests into sensitive or insensitive classes to latency, and thereafter suitable VM can be selected for migration.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cloud computing service models have experienced rapid growth and inefficient
resource usage is known as one of the greatest causes of high energy
consumption in cloud data centers. Resource allocation in cloud data centers
aiming to reduce energy consumption has been conducted using live migration of
Virtual Machines (VMs) and their consolidation into the small number of
Physical Machines (PMs). However, the selection of the appropriate VM for
migration is an important challenge. To solve this issue, VMs can be classified
according to the pattern of user requests into sensitive or insensitive classes
to latency, and thereafter suitable VMs can be selected for migration. In this
paper, the combination of Convolution Neural Network (CNN) and Gated Recurrent
Unit (GRU) is utilized for the classification of VMs in the Microsoft Azure
dataset. Due to the fact the majority of VMs in this dataset are labeled as
insensitive to latency, migration of more VMs in this group not only reduces
energy consumption but also decreases the violation of Service Level Agreements
(SLA). Based on the empirical results, the proposed model obtained an accuracy
of 95.18which clearly demonstrates the superiority of our proposed model
compared to other existing models.
Related papers
- Enhancing Robustness and Efficiency of Least Square Twin SVM via Granular Computing [0.2999888908665658]
In the domain of machine learning, least square twin support vector machine (LSTSVM) stands out as one of the state-of-the-art models.
LSTSVM suffers from sensitivity to noise and inversions, overlooking the principle and instability in resampling.
We propose the robust granular ball LSTSVM (GBLSTSVM), which is trained using granular balls instead of original data points.
arXiv Detail & Related papers (2024-10-22T18:13:01Z) - Ditto: Elastic Confidential VMs with Secure and Dynamic CPU Scaling [35.971391128345125]
"Elastic CVM" and the Worker vCPU design pave the way for more flexible and cost-effective confidential computing environments.
"Elastic CVM" and the Worker vCPU design not only optimize cloud resource utilization but also pave the way for more flexible and cost-effective confidential computing environments.
arXiv Detail & Related papers (2024-09-23T20:52:10Z) - Hybrid Heterogeneous Clusters Can Lower the Energy Consumption of LLM Inference Workloads [0.2389598109913753]
Training and using Large Language Models (LLMs) require large amounts of energy.
This paper addresses the challenge of reducing energy consumption in data centers running LLMs.
We propose a hybrid data center model that uses a cost-based scheduling framework to dynamically allocate tasks across hardware accelerators.
arXiv Detail & Related papers (2024-04-25T11:24:08Z) - HyperVQ: MLR-based Vector Quantization in Hyperbolic Space [56.4245885674567]
We study the use of hyperbolic spaces for vector quantization (HyperVQ)
We show that hyperVQ performs comparably in reconstruction and generative tasks while outperforming VQ in discriminative tasks and learning a highly disentangled latent space.
arXiv Detail & Related papers (2024-03-18T03:17:08Z) - SpotServe: Serving Generative Large Language Models on Preemptible
Instances [64.18638174004151]
SpotServe is the first distributed large language models serving system on preemptible instances.
We show that SpotServe can reduce the P99 tail latency by 2.4 - 9.1x compared with the best existing LLM serving systems.
We also show that SpotServe can leverage the price advantage of preemptive instances, saving 54% monetary cost compared with only using on-demand instances.
arXiv Detail & Related papers (2023-11-27T06:31:17Z) - POLCA: Power Oversubscription in LLM Cloud Providers [0.8299593158757622]
Large language models (LLMs) are becoming increasingly power intensive.
We show that there is a significant opportunity to oversubscribe power in LLM clusters.
We propose POLCA, our framework for power oversubscription that is robust, reliable, and readily deployable for GPU clusters.
arXiv Detail & Related papers (2023-08-24T16:32:34Z) - Adaptive Federated Pruning in Hierarchical Wireless Networks [69.6417645730093]
Federated Learning (FL) is a privacy-preserving distributed learning framework where a server aggregates models updated by multiple devices without accessing their private datasets.
In this paper, we introduce model pruning for HFL in wireless networks to reduce the neural network scale.
We show that our proposed HFL with model pruning achieves similar learning accuracy compared with the HFL without model pruning and reduces about 50 percent communication cost.
arXiv Detail & Related papers (2023-05-15T22:04:49Z) - Energy-Efficient Model Compression and Splitting for Collaborative
Inference Over Time-Varying Channels [52.60092598312894]
We propose a technique to reduce the total energy bill at the edge device by utilizing model compression and time-varying model split between the edge and remote nodes.
Our proposed solution results in minimal energy consumption and $CO$ emission compared to the considered baselines.
arXiv Detail & Related papers (2021-06-02T07:36:27Z) - A Machine Learning-Based Migration Strategy for Virtual Network Function
Instances [3.7783523378336104]
We develop the VNF Neural Network for Instance Migration (VNNIM), a migration strategy for VNF instances.
VNNIM is very effective in predicting the post-migration server exhibiting a binary accuracy of 99.07%.
The greatest advantage of VNNIM, however, is its run-time efficiency highlighted through a run-time analysis.
arXiv Detail & Related papers (2020-06-15T15:03:27Z) - Communication Efficient Federated Learning with Energy Awareness over
Wireless Networks [51.645564534597625]
In federated learning (FL), the parameter server and the mobile devices share the training parameters over wireless links.
We adopt the idea of SignSGD in which only the signs of the gradients are exchanged.
Two optimization problems are formulated and solved, which optimize the learning performance.
Considering that the data may be distributed across the mobile devices in a highly uneven fashion in FL, a sign-based algorithm is proposed.
arXiv Detail & Related papers (2020-04-15T21:25:13Z) - On Coresets for Support Vector Machines [61.928187390362176]
A coreset is a small, representative subset of the original data points.
We show that our algorithm can be used to extend the applicability of any off-the-shelf SVM solver to streaming, distributed, and dynamic data settings.
arXiv Detail & Related papers (2020-02-15T23:25:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.