Caching Placement and Resource Allocation for Cache-Enabling UAV NOMA
Networks
- URL: http://arxiv.org/abs/2008.05168v1
- Date: Wed, 12 Aug 2020 08:33:51 GMT
- Title: Caching Placement and Resource Allocation for Cache-Enabling UAV NOMA
Networks
- Authors: Tiankui Zhang, Ziduan Wang, Yuanwei Liu, Wenjun Xu and Arumugam
Nallanathan
- Abstract summary: This article investigates the cache-enabling unmanned aerial vehicle (UAV) cellular networks with massive access capability supported by non-orthogonal multiple access (NOMA)
We formulate the long-term caching placement and resource allocation optimization problem for content delivery delay minimization as a Markov decision process (MDP)
We propose a Q-learning based caching placement and resource allocation algorithm, where the UAV learns and selects action with emphsoft $varepsilon$-greedy strategy to search for the optimal match between actions and states.
- Score: 87.6031308969681
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This article investigates the cache-enabling unmanned aerial vehicle (UAV)
cellular networks with massive access capability supported by non-orthogonal
multiple access (NOMA). The delivery of a large volume of multimedia contents
for ground users is assisted by a mobile UAV base station, which caches some
popular contents for wireless backhaul link traffic offloading. In
cache-enabling UAV NOMA networks, the caching placement of content caching
phase and radio resource allocation of content delivery phase are crucial for
network performance. To cope with the dynamic UAV locations and content
requests in practical scenarios, we formulate the long-term caching placement
and resource allocation optimization problem for content delivery delay
minimization as a Markov decision process (MDP). The UAV acts as an agent to
take actions for caching placement and resource allocation, which includes the
user scheduling of content requests and the power allocation of NOMA users. In
order to tackle the MDP, we propose a Q-learning based caching placement and
resource allocation algorithm, where the UAV learns and selects action with
\emph{soft ${\varepsilon}$-greedy} strategy to search for the optimal match
between actions and states. Since the action-state table size of Q-learning
grows with the number of states in the dynamic networks, we propose a function
approximation based algorithm with combination of stochastic gradient descent
and deep neural networks, which is suitable for large-scale networks. Finally,
the numerical results show that the proposed algorithms provide considerable
performance compared to benchmark algorithms, and obtain a trade-off between
network performance and calculation complexity.
Related papers
- Cluster-Based Multi-Agent Task Scheduling for Space-Air-Ground Integrated Networks [60.085771314013044]
Low-altitude economy holds significant potential for development in areas such as communication and sensing.
We propose a Clustering-based Multi-agent Deep Deterministic Policy Gradient (CMADDPG) algorithm to address the multi-UAV cooperative task scheduling challenges in SAGIN.
arXiv Detail & Related papers (2024-12-14T06:17:33Z) - IoV Scenario: Implementation of a Bandwidth Aware Algorithm in Wireless
Network Communication Mode [49.734868032441625]
This paper proposes a bandwidth aware multi domain virtual network embedding algorithm (BA-VNE)
The algorithm is mainly aimed at the problem that users need a lot of bandwidth in wireless communication mode.
In order to improve the performance of the algorithm, we introduce particle swarm optimization (PSO) algorithm.
arXiv Detail & Related papers (2022-02-03T03:34:06Z) - Multi-Task Network Pruning and Embedded Optimization for Real-time
Deployment in ADAS [0.0]
Camera-based Deep Learning algorithms are increasingly needed for perception in Automated Driving systems.
constraints from the automotive industry challenge the deployment of CNNs by imposing embedded systems with limited computational resources.
We propose an approach to embed a multi-task CNN network under such conditions on a commercial prototype platform.
arXiv Detail & Related papers (2021-01-19T19:29:38Z) - Multi-Agent Reinforcement Learning in NOMA-aided UAV Networks for
Cellular Offloading [59.32570888309133]
A novel framework is proposed for cellular offloading with the aid of multiple unmanned aerial vehicles (UAVs)
Non-orthogonal multiple access (NOMA) technique is employed at each UAV to further improve the spectrum efficiency of the wireless network.
A mutual deep Q-network (MDQN) algorithm is proposed to jointly determine the optimal 3D trajectory and power allocation of UAVs.
arXiv Detail & Related papers (2020-10-18T20:22:05Z) - Deep Reinforcement Learning Based Mobile Edge Computing for Intelligent
Internet of Things [10.157016543999045]
We devise the system by proposing the offloading strategy intelligently through the deep reinforcement learning algorithm.
Deep Q-Network is used to automatically learn the offloading decision in order to optimize the system performance.
A neural network (NN) is trained to predict the offloading action, where the training data is generated from the environmental system.
In particular, the system cost of latency and energy consumption can be reduced significantly by the proposed deep reinforcement learning based algorithm.
arXiv Detail & Related papers (2020-08-01T11:45:54Z) - Computation Offloading in Multi-Access Edge Computing Networks: A
Multi-Task Learning Approach [7.203439085947118]
Multi-access edge computing (MEC) has already shown the potential in enabling mobile devices to bear the computation-intensive applications by offloading some tasks to a nearby access point (AP) integrated with a MEC server (MES)
However, due to the varying network conditions and limited computation resources of the MES, the offloading decisions taken by a mobile device and the computational resources allocated by the MES may not be efficiently achieved with the lowest cost.
We propose a dynamic offloading framework for the MEC network, in which the uplink non-orthogonal multiple access (NOMA) is used to enable multiple devices to upload their
arXiv Detail & Related papers (2020-06-29T15:11:10Z) - Reinforcement Learning for Caching with Space-Time Popularity Dynamics [61.55827760294755]
caching is envisioned to play a critical role in next-generation networks.
To intelligently prefetch and store contents, a cache node should be able to learn what and when to cache.
This chapter presents a versatile reinforcement learning based approach for near-optimal caching policy design.
arXiv Detail & Related papers (2020-05-19T01:23:51Z) - A Non-Stationary Bandit-Learning Approach to Energy-Efficient
Femto-Caching with Rateless-Coded Transmission [98.47527781626161]
We study a resource allocation problem for joint caching and transmission in small cell networks.
We then formulate the problem as to select a file from the cache together with a transmission power level for every broadcast round.
In contrast to the state-of-the-art research, the proposed approach is especially suitable for networks with time-variant statistical properties.
arXiv Detail & Related papers (2020-04-13T09:07:17Z) - Reinforcement Learning Based Cooperative Coded Caching under Dynamic
Popularities in Ultra-Dense Networks [38.44125997148742]
caching strategy at small base stations (SBSs) is critical to meet massive high data rate requests.
We exploit reinforcement learning (RL) to design a cooperative caching strategy with maximum-distance separable (MDS) coding.
arXiv Detail & Related papers (2020-03-08T10:45:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.