Handover and SINR-Aware Path Optimization in 5G-UAV mmWave Communication using DRL
- URL: http://arxiv.org/abs/2504.02688v1
- Date: Thu, 03 Apr 2025 15:28:04 GMT
- Title: Handover and SINR-Aware Path Optimization in 5G-UAV mmWave Communication using DRL
- Authors: Achilles Kiwanuka Machumilane, Alberto Gotta, Pietro CassarĂ ,
- Abstract summary: We propose a novel model-free actor-critic deep reinforcement learning (AC-DRL) framework for path optimization in UAV-assisted 5G mmWave wireless networks.<n>We train an AC-RL agent that enables a UAV connected to a gNB to determine the optimal path to a desired destination in the shortest possible time.
- Score: 0.5315148938765306
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Path planning and optimization for unmanned aerial vehicles (UAVs)-assisted next-generation wireless networks is critical for mobility management and ensuring UAV safety and ubiquitous connectivity, especially in dense urban environments with street canyons and tall buildings. Traditional statistical and model-based techniques have been successfully used for path optimization in communication networks. However, when dynamic channel propagation characteristics such as line-of-sight (LOS), interference, handover, and signal-to-interference and noise ratio (SINR) are included in path optimization, statistical and model-based path planning solutions become obsolete since they cannot adapt to the dynamic and time-varying wireless channels, especially in the mmWave bands. In this paper, we propose a novel model-free actor-critic deep reinforcement learning (AC-DRL) framework for path optimization in UAV-assisted 5G mmWave wireless networks, which combines four important aspects of UAV communication: \textit{flight time, handover, connectivity and SINR}. We train an AC-RL agent that enables a UAV connected to a gNB to determine the optimal path to a desired destination in the shortest possible time with minimal gNB handover, while maintaining connectivity and the highest possible SINR. We train our model with data from a powerful ray tracing tool called Wireless InSite, which uses 3D images of the propagation environment and provides data that closely resembles the real propagation environment. The simulation results show that our system has superior performance in tracking high SINR compared to other selected RL algorithms.
Related papers
- AoI-Sensitive Data Forwarding with Distributed Beamforming in UAV-Assisted IoT [32.6091251316091]
This paper proposes a UAV-assisted system based on distributed beamforming to enhance age forwarding information (AoI) in Internet of Things (IoT)<n>We propose a deep reinforcement learning (DRL)-based algorithm to solve the problem, thereby enhancing stability and accelerate convergence.
arXiv Detail & Related papers (2025-02-13T07:48:36Z) - DRL Optimization Trajectory Generation via Wireless Network Intent-Guided Diffusion Models for Optimizing Resource Allocation [58.62766376631344]
We propose a customized wireless network intent (WNI-G) model to address different state variations of wireless communication networks.
Extensive simulation achieves greater stability in spectral efficiency and variations of traditional DRL models in dynamic communication systems.
arXiv Detail & Related papers (2024-10-18T14:04:38Z) - Joint Optimization of Deployment and Trajectory in UAV and IRS-Assisted
IoT Data Collection System [25.32139119893323]
Unmanned aerial vehicles (UAVs) can be applied in many Internet of Things (IoT) systems.
The UAV-IoT wireless channels may be occasionally blocked by trees or high-rise buildings.
This article aims to minimize the energy consumption of the system by jointly optimizing the deployment and trajectory of the UAV.
arXiv Detail & Related papers (2022-10-27T06:27:40Z) - Trajectory Design for UAV-Based Internet-of-Things Data Collection: A
Deep Reinforcement Learning Approach [93.67588414950656]
In this paper, we investigate an unmanned aerial vehicle (UAV)-assisted Internet-of-Things (IoT) system in a 3D environment.
We present a TD3-based trajectory design for completion time minimization (TD3-TDCTM) algorithm.
Our simulation results show the superiority of the proposed TD3-TDCTM algorithm over three conventional non-learning based baseline methods.
arXiv Detail & Related papers (2021-07-23T03:33:29Z) - Model-aided Deep Reinforcement Learning for Sample-efficient UAV
Trajectory Design in IoT Networks [20.303937220315177]
We propose a model-aided deep Q-learning approach to guide a flight-time restricted UAV on a data harvesting mission.
We show that in comparison with standard DRL approaches, the proposed model-aided approach requires at least one order of magnitude less training data samples to reach identical data collection performance.
arXiv Detail & Related papers (2021-04-21T08:25:11Z) - Jamming-Resilient Path Planning for Multiple UAVs via Deep Reinforcement
Learning [1.2330326247154968]
Unmanned aerial vehicles (UAVs) are expected to be an integral part of wireless networks.
In this paper, we aim to find collision-free paths for multiple cellular-connected UAVs.
We propose an offline temporal difference (TD) learning algorithm with online signal-to-interference-plus-noise ratio mapping to solve the problem.
arXiv Detail & Related papers (2021-04-09T16:52:33Z) - Learning-Based UAV Trajectory Optimization with Collision Avoidance and
Connectivity Constraints [0.0]
Unmanned aerial vehicles (UAVs) are expected to be an integral part of wireless networks.
In this paper, we reformulate the multi-UAV trajectory optimization problem with collision avoidance and wireless connectivity constraints.
We propose a decentralized deep reinforcement learning approach to solve the problem.
arXiv Detail & Related papers (2021-04-03T22:22:20Z) - Distributed Conditional Generative Adversarial Networks (GANs) for
Data-Driven Millimeter Wave Communications in UAV Networks [116.94802388688653]
A novel framework is proposed to perform data-driven air-to-ground (A2G) channel estimation for millimeter wave (mmWave) communications in an unmanned aerial vehicle (UAV) wireless network.
An effective channel estimation approach is developed, allowing each UAV to train a stand-alone channel model via a conditional generative adversarial network (CGAN) along each beamforming direction.
A cooperative framework, based on a distributed CGAN architecture, is developed, allowing each UAV to collaboratively learn the mmWave channel distribution.
arXiv Detail & Related papers (2021-02-02T20:56:46Z) - Distributed Multi-agent Meta Learning for Trajectory Design in Wireless
Drone Networks [151.27147513363502]
This paper studies the problem of the trajectory design for a group of energyconstrained drones operating in dynamic wireless network environments.
A value based reinforcement learning (VDRL) solution and a metatraining mechanism is proposed.
arXiv Detail & Related papers (2020-12-06T01:30:12Z) - Distributional Reinforcement Learning for mmWave Communications with
Intelligent Reflectors on a UAV [119.97450366894718]
A novel communication framework that uses an unmanned aerial vehicle (UAV)-carried intelligent reflector (IR) is proposed.
In order to maximize the downlink sum-rate, the optimal precoding matrix (at the base station) and reflection coefficient (at the IR) are jointly derived.
arXiv Detail & Related papers (2020-11-03T16:50:37Z) - Data Freshness and Energy-Efficient UAV Navigation Optimization: A Deep
Reinforcement Learning Approach [88.45509934702913]
We design a navigation policy for multiple unmanned aerial vehicles (UAVs) where mobile base stations (BSs) are deployed.
We incorporate different contextual information such as energy and age of information (AoI) constraints to ensure the data freshness at the ground BS.
By applying the proposed trained model, an effective real-time trajectory policy for the UAV-BSs captures the observable network states over time.
arXiv Detail & Related papers (2020-02-21T07:29:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.