VMAgent: Scheduling Simulator for Reinforcement Learning
- URL: http://arxiv.org/abs/2112.04785v1
- Date: Thu, 9 Dec 2021 09:18:38 GMT
- Title: VMAgent: Scheduling Simulator for Reinforcement Learning
- Authors: Junjie Sheng and Shengliang Cai and Haochuan Cui and Wenhao Li and Yun
Hua and Bo Jin and Wenli Zhou and Yiqiu Hu and Lei Zhu and Qian Peng and
Hongyuan Zha and Xiangfeng Wang
- Abstract summary: A novel simulator called VMAgent is introduced to help RL researchers better explore new methods.
VMAgent is inspired by practical virtual machine (VM) scheduling tasks.
From the VM scheduling perspective, VMAgent also helps to explore better learning-based scheduling solutions.
- Score: 44.026076801936874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A novel simulator called VMAgent is introduced to help RL researchers better
explore new methods, especially for virtual machine scheduling. VMAgent is
inspired by practical virtual machine (VM) scheduling tasks and provides an
efficient simulation platform that can reflect the real situations of cloud
computing. Three scenarios (fading, recovering, and expansion) are concluded
from practical cloud computing and corresponds to many reinforcement learning
challenges (high dimensional state and action spaces, high non-stationarity,
and life-long demand). VMAgent provides flexible configurations for RL
researchers to design their customized scheduling environments considering
different problem features. From the VM scheduling perspective, VMAgent also
helps to explore better learning-based scheduling solutions.
Related papers
- Cost-Aware Dynamic Cloud Workflow Scheduling using Self-Attention and Evolutionary Reinforcement Learning [7.653021685451039]
This paper proposes a novel self-attention policy network for cloud workflow scheduling.
The trained SPN-CWS can effectively process all candidate instances simultaneously to identify the most suitable VM instance to execute every workflow task.
Our method can noticeably outperform several state-of-the-art algorithms on multiple benchmark CDMWS problems.
arXiv Detail & Related papers (2024-09-27T04:45:06Z) - Waymax: An Accelerated, Data-Driven Simulator for Large-Scale Autonomous
Driving Research [76.93956925360638]
Waymax is a new data-driven simulator for autonomous driving in multi-agent scenes.
It runs entirely on hardware accelerators such as TPUs/GPUs and supports in-graph simulation for training.
We benchmark a suite of popular imitation and reinforcement learning algorithms with ablation studies on different design decisions.
arXiv Detail & Related papers (2023-10-12T20:49:15Z) - In Situ Framework for Coupling Simulation and Machine Learning with
Application to CFD [51.04126395480625]
Recent years have seen many successful applications of machine learning (ML) to facilitate fluid dynamic computations.
As simulations grow, generating new training datasets for traditional offline learning creates I/O and storage bottlenecks.
This work offers a solution by simplifying this coupling and enabling in situ training and inference on heterogeneous clusters.
arXiv Detail & Related papers (2023-06-22T14:07:54Z) - Sim2real Transfer Learning for Point Cloud Segmentation: An Industrial
Application Case on Autonomous Disassembly [55.41644538483948]
We present an industrial application case that uses sim2real transfer learning for point cloud data.
We provide insights on how to generate and process synthetic point cloud data.
A novel patch-based attention network is proposed additionally to tackle this problem.
arXiv Detail & Related papers (2023-01-12T14:00:37Z) - Parallel Reinforcement Learning Simulation for Visual Quadrotor
Navigation [4.597465975849579]
Reinforcement learning (RL) is an agent-based approach for teaching robots to navigate within the physical world.
We present a simulation framework, built on AirSim, which provides efficient parallel training.
Building on this framework, Ape-X is modified to incorporate decentralised training of AirSim environments.
arXiv Detail & Related papers (2022-09-22T15:27:42Z) - SOLIS -- The MLOps journey from data acquisition to actionable insights [62.997667081978825]
In this paper we present a unified deployment pipeline and freedom-to-operate approach that supports all requirements while using basic cross-platform tensor framework and script language engines.
This approach however does not supply the needed procedures and pipelines for the actual deployment of machine learning capabilities in real production grade systems.
arXiv Detail & Related papers (2021-12-22T14:45:37Z) - Molecular Dynamics Simulations on Cloud Computing and Machine Learning
Platforms [0.8093262393618671]
We see a paradigm shift in the computational structure, design, and requirements of scientific computing applications.
Data-driven and machine learning approaches are being used to support, speed-up, and enhance scientific computing applications.
Cloud computing platforms are increasingly appealing for scientific computing.
arXiv Detail & Related papers (2021-11-11T21:20:26Z) - Multitask Adaptation by Retrospective Exploration with Learned World
Models [77.34726150561087]
We propose a meta-learned addressing model called RAMa that provides training samples for the MBRL agent taken from task-agnostic storage.
The model is trained to maximize the expected agent's performance by selecting promising trajectories solving prior tasks from the storage.
arXiv Detail & Related papers (2021-10-25T20:02:57Z) - Multi-factorial Optimization for Large-scale Virtual Machine Placement
in Cloud Computing [15.840835256016259]
Evolutionary algorithms (EAs) have been performed promising-solving on virtual machine placement (VMP) problems in the past.
As growing demand for cloud services, the existing EAs fail to implement in large-scale virtual machine placement problem.
This paper aims to apply the MFO technology to the LVMP problem in heterogeneous environment.
arXiv Detail & Related papers (2020-01-18T02:59:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.