A Bayesian Framework of Deep Reinforcement Learning for Joint O-RAN/MEC
Orchestration
- URL: http://arxiv.org/abs/2312.16142v1
- Date: Tue, 26 Dec 2023 18:04:49 GMT
- Title: A Bayesian Framework of Deep Reinforcement Learning for Joint O-RAN/MEC
Orchestration
- Authors: Fahri Wisnu Murti, Samad Ali, Matti Latva-aho
- Abstract summary: Multi-access Edge Computing (MEC) can be implemented together with Open Radio Access Network (O-RAN) over commodity platforms to offer low-cost deployment.
In this paper, a joint O-RAN/MEC orchestration using a Bayesian deep reinforcement learning (RL)-based framework is proposed.
- Score: 12.914011030970814
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-access Edge Computing (MEC) can be implemented together with Open Radio
Access Network (O-RAN) over commodity platforms to offer low-cost deployment
and bring the services closer to end-users. In this paper, a joint O-RAN/MEC
orchestration using a Bayesian deep reinforcement learning (RL)-based framework
is proposed that jointly controls the O-RAN functional splits, the allocated
resources and hosting locations of the O-RAN/MEC services across
geo-distributed platforms, and the routing for each O-RAN/MEC data flow. The
goal is to minimize the long-term overall network operation cost and maximize
the MEC performance criterion while adapting possibly time-varying O-RAN/MEC
demands and resource availability. This orchestration problem is formulated as
Markov decision process (MDP). However, the system consists of multiple BSs
that share the same resources and serve heterogeneous demands, where their
parameters have non-trivial relations. Consequently, finding the exact model of
the underlying system is impractical, and the formulated MDP renders in a large
state space with multi-dimensional discrete action. To address such modeling
and dimensionality issues, a novel model-free RL agent is proposed for our
solution framework. The agent is built from Double Deep Q-network (DDQN) that
tackles the large state space and is then incorporated with action branching,
an action decomposition method that effectively addresses the multi-dimensional
discrete action with linear increase complexity. Further, an efficient
exploration-exploitation strategy under a Bayesian framework using Thomson
sampling is proposed to improve the learning performance and expedite its
convergence. Trace-driven simulations are performed using an O-RAN-compliant
model. The results show that our approach is data-efficient (i.e., converges
faster) and increases the returned reward by 32\% than its non-Bayesian
version.
Related papers
- FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - Efficient Multi-agent Reinforcement Learning by Planning [33.51282615335009]
Multi-agent reinforcement learning (MARL) algorithms have accomplished remarkable breakthroughs in solving large-scale decision-making tasks.
Most existing MARL algorithms are model-free, limiting sample efficiency and hindering their applicability in more challenging scenarios.
We propose the MAZero algorithm, which combines a centralized model with Monte Carlo Tree Search (MCTS) for policy search.
arXiv Detail & Related papers (2024-05-20T04:36:02Z) - A Multi-Head Ensemble Multi-Task Learning Approach for Dynamical
Computation Offloading [62.34538208323411]
We propose a multi-head ensemble multi-task learning (MEMTL) approach with a shared backbone and multiple prediction heads (PHs)
MEMTL outperforms benchmark methods in both the inference accuracy and mean square error without requiring additional training data.
arXiv Detail & Related papers (2023-09-02T11:01:16Z) - Maximize to Explore: One Objective Function Fusing Estimation, Planning,
and Exploration [87.53543137162488]
We propose an easy-to-implement online reinforcement learning (online RL) framework called textttMEX.
textttMEX integrates estimation and planning components while balancing exploration exploitation automatically.
It can outperform baselines by a stable margin in various MuJoCo environments with sparse rewards.
arXiv Detail & Related papers (2023-05-29T17:25:26Z) - A Multi-objective Complex Network Pruning Framework Based on
Divide-and-conquer and Global Performance Impairment Ranking [40.59001171151929]
A multi-objective complex network pruning framework based on divide-and-conquer and global performance impairment ranking is proposed in this paper.
The proposed algorithm achieves a comparable performance with the state-of-the-art pruning methods.
arXiv Detail & Related papers (2023-03-28T12:05:15Z) - Adaptive Stochastic ADMM for Decentralized Reinforcement Learning in
Edge Industrial IoT [106.83952081124195]
Reinforcement learning (RL) has been widely investigated and shown to be a promising solution for decision-making and optimal control processes.
We propose an adaptive ADMM (asI-ADMM) algorithm and apply it to decentralized RL with edge-computing-empowered IIoT networks.
Experiment results show that our proposed algorithms outperform the state of the art in terms of communication costs and scalability, and can well adapt to complex IoT environments.
arXiv Detail & Related papers (2021-06-30T16:49:07Z) - Deep Multimodal Fusion by Channel Exchanging [87.40768169300898]
This paper proposes a parameter-free multimodal fusion framework that dynamically exchanges channels between sub-networks of different modalities.
The validity of such exchanging process is also guaranteed by sharing convolutional filters yet keeping separate BN layers across modalities, which, as an add-on benefit, allows our multimodal architecture to be almost as compact as a unimodal network.
arXiv Detail & Related papers (2020-11-10T09:53:20Z) - Coded Stochastic ADMM for Decentralized Consensus Optimization with Edge
Computing [113.52575069030192]
Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones and vehicles.
Due to the limitations of communication costs and security requirements, it is of paramount importance to extract information in a decentralized manner instead of aggregating data to a fusion center.
We consider the problem of learning model parameters in a multi-agent system with data locally processed via distributed edge nodes.
A class of mini-batch alternating direction method of multipliers (ADMM) algorithms is explored to develop the distributed learning model.
arXiv Detail & Related papers (2020-10-02T10:41:59Z) - Distributed Resource Scheduling for Large-Scale MEC Systems: A
Multi-Agent Ensemble Deep Reinforcement Learning with Imitation Acceleration [44.40722828581203]
We propose a distributed intelligent resource scheduling (DIRS) framework, which includes centralized training relying on the global information and distributed decision making by each agent deployed in each MEC server.
We first introduce a novel multi-agent ensemble-assisted distributed deep reinforcement learning (DRL) architecture, which can simplify the overall neural network structure of each agent.
Secondly, we apply action refinement to enhance the exploration ability of the proposed DIRS framework, where the near-optimal state-action pairs are obtained by a novel L'evy flight search.
arXiv Detail & Related papers (2020-05-21T20:04:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.