V2N Service Scaling with Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2301.13324v2
- Date: Wed, 1 Feb 2023 01:51:09 GMT
- Title: V2N Service Scaling with Deep Reinforcement Learning
- Authors: Cyril Shih-Huan Hsu, Jorge Mart\'in-P\'erez, Chrysa Papagianni, Paola
Grosso
- Abstract summary: We employ Deep Reinforcement Learning (DRL) for vertical scaling in Edge computing to support vehicular-to-network communications.
We show that DDPG outperforms existing solutions, reducing the average number of active CPUs by 23% while increasing the long-term reward by 24%.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The fifth generation (5G) of wireless networks is set out to meet the
stringent requirements of vehicular use cases. Edge computing resources can aid
in this direction by moving processing closer to end-users, reducing latency.
However, given the stochastic nature of traffic loads and availability of
physical resources, appropriate auto-scaling mechanisms need to be employed to
support cost-efficient and performant services. To this end, we employ Deep
Reinforcement Learning (DRL) for vertical scaling in Edge computing to support
vehicular-to-network communications. We address the problem using Deep
Deterministic Policy Gradient (DDPG). As DDPG is a model-free off-policy
algorithm for learning continuous actions, we introduce a discretization
approach to support discrete scaling actions. Thus we address scalability
problems inherent to high-dimensional discrete action spaces. Employing a
real-world vehicular trace data set, we show that DDPG outperforms existing
solutions, reducing (at minimum) the average number of active CPUs by 23% while
increasing the long-term reward by 24%.
Related papers
- Two-Timescale Model Caching and Resource Allocation for Edge-Enabled AI-Generated Content Services [55.0337199834612]
Generative AI (GenAI) has emerged as a transformative technology, enabling customized and personalized AI-generated content (AIGC) services.
These services require executing GenAI models with billions of parameters, posing significant obstacles to resource-limited wireless edge.
We introduce the formulation of joint model caching and resource allocation for AIGC services to balance a trade-off between AIGC quality and latency metrics.
arXiv Detail & Related papers (2024-11-03T07:01:13Z) - FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - DNN Partitioning, Task Offloading, and Resource Allocation in Dynamic Vehicular Networks: A Lyapunov-Guided Diffusion-Based Reinforcement Learning Approach [49.56404236394601]
We formulate the problem of joint DNN partitioning, task offloading, and resource allocation in Vehicular Edge Computing.
Our objective is to minimize the DNN-based task completion time while guaranteeing the system stability over time.
We propose a Multi-Agent Diffusion-based Deep Reinforcement Learning (MAD2RL) algorithm, incorporating the innovative use of diffusion models.
arXiv Detail & Related papers (2024-06-11T06:31:03Z) - DaCapo: Accelerating Continuous Learning in Autonomous Systems for Video Analytics [4.962007210373803]
We propose a hardware-algorithm co-designed solution for continuous learning, DaCapo, that enables autonomous systems to perform concurrent executions of inference, labeling, and training.
DaCapo achieves 6.5% and 5.5% higher accuracy than a state-of-the-art continuous learning systems, Ekya and EOMU, respectively, while consuming 254x less power.
arXiv Detail & Related papers (2024-03-21T12:28:44Z) - A Deep RL Approach on Task Placement and Scaling of Edge Resources for Cellular Vehicle-to-Network Service Provisioning [6.625994697789603]
We tackle the interdependent problems of service task placement and scaling of edge resources.
We introduce a Deep Hybrid Policy Gradient (DHPG), a Deep Reinforcement Learning (DRL) approach for hybrid action spaces.
The performance of DHPG is evaluated against several state-of-the-art (SoA) solutions through simulations employing a real-world C-V2N traffic dataset.
arXiv Detail & Related papers (2023-05-16T22:19:19Z) - Graph Reinforcement Learning for Radio Resource Allocation [13.290246410488727]
We resort to graph reinforcement learning for exploiting two kinds of relational priors inherent in many problems in wireless communications.
To design graph reinforcement learning framework systematically, we first conceive a method to transform state matrix into state graph.
We then propose a general method for graph neural networks to satisfy desirable permutation properties.
arXiv Detail & Related papers (2022-03-08T08:02:54Z) - Multi-Exit Semantic Segmentation Networks [78.44441236864057]
We propose a framework for converting state-of-the-art segmentation models to MESS networks.
specially trained CNNs that employ parametrised early exits along their depth to save during inference on easier samples.
We co-optimise the number, placement and architecture of the attached segmentation heads, along with the exit policy, to adapt to the device capabilities and application-specific requirements.
arXiv Detail & Related papers (2021-06-07T11:37:03Z) - Learning to Optimize Industry-Scale Dynamic Pickup and Delivery Problems [17.076557377480444]
The Dynamic Pickup and Delivery Problem (D PDP) is aimed at dynamically scheduling vehicles among multiple sites in order to minimize the cost when delivery orders are not known a priori.
We propose a data-driven approach, Spatial-Temporal Aided Double Deep Graph Network (ST-DDGN), to solve industry-scale D PDP.
Our method is entirely data driven and thus adaptive, i.e., the relational representation of adjacent vehicles can be learned and corrected by ST-DDGN from data periodically.
arXiv Detail & Related papers (2021-05-27T01:16:00Z) - Adaptive Subcarrier, Parameter, and Power Allocation for Partitioned
Edge Learning Over Broadband Channels [69.18343801164741]
partitioned edge learning (PARTEL) implements parameter-server training, a well known distributed learning method, in wireless network.
We consider the case of deep neural network (DNN) models which can be trained using PARTEL by introducing some auxiliary variables.
arXiv Detail & Related papers (2020-10-08T15:27:50Z) - When Deep Reinforcement Learning Meets Federated Learning: Intelligent
Multi-Timescale Resource Management for Multi-access Edge Computing in 5G
Ultra Dense Network [31.274279003934268]
We first propose an intelligent ultra-dense edge computing (I-UDEC) framework, which integrates blockchain and AI into 5G edge computing networks.
In order to achieve real-time and low overhead computation offloading decisions and resource allocation strategies, we design a novel two-timescale deep reinforcement learning (textit2Ts-DRL) approach.
Our proposed algorithm can reduce task execution time up to 31.87%.
arXiv Detail & Related papers (2020-09-22T15:08:00Z) - Joint Parameter-and-Bandwidth Allocation for Improving the Efficiency of
Partitioned Edge Learning [73.82875010696849]
Machine learning algorithms are deployed at the network edge for training artificial intelligence (AI) models.
This paper focuses on the novel joint design of parameter (computation load) allocation and bandwidth allocation.
arXiv Detail & Related papers (2020-03-10T05:52:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.