GCN-Driven Reinforcement Learning for Probabilistic Real-Time Guarantees in Industrial URLLC
- URL: http://arxiv.org/abs/2506.15011v2
- Date: Mon, 07 Jul 2025 20:38:38 GMT
- Title: GCN-Driven Reinforcement Learning for Probabilistic Real-Time Guarantees in Industrial URLLC
- Authors: Eman Alqudah, Ashfaq Khokhar,
- Abstract summary: We enhance the Local Deadline Partition (LDP) algorithm by introducing a Graph Convolutional Network (GCN) integrated with a Deep Q-Network (DQN) reinforcement learning framework.<n>We show that our GCN-DQN model achieves mean SINR improvements of 179.6%, 197.4%, and 175.2% over LDP across three network configurations.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Ensuring packet-level communication quality is vital for ultra-reliable, low-latency communications (URLLC) in large-scale industrial wireless networks. We enhance the Local Deadline Partition (LDP) algorithm by introducing a Graph Convolutional Network (GCN) integrated with a Deep Q-Network (DQN) reinforcement learning framework for improved interference coordination in multi-cell, multi-channel networks. Unlike LDP's static priorities, our approach dynamically learns link priorities based on real-time traffic demand, network topology, remaining transmission opportunities, and interference patterns. The GCN captures spatial dependencies, while the DQN enables adaptive scheduling decisions through reward-guided exploration. Simulation results show that our GCN-DQN model achieves mean SINR improvements of 179.6\%, 197.4\%, and 175.2\% over LDP across three network configurations. Additionally, the GCN-DQN model demonstrates mean SINR improvements of 31.5\%, 53.0\%, and 84.7\% over our previous CNN-based approach across the same configurations. These results underscore the effectiveness of our GCN-DQN model in addressing complex URLLC requirements with minimal overhead and superior network performance.
Related papers
- CNN-Enabled Scheduling for Probabilistic Real-Time Guarantees in Industrial URLLC [0.0]
We introduce a CNN-based dynamic priority prediction mechanism for improved interference coordination in multi-cell, multi-channel networks.<n>We show SINR gains of up to 113%, 94%, and 49% over LDP across three network configurations.
arXiv Detail & Related papers (2025-06-17T21:40:19Z) - Opportunistic Routing in Wireless Communications via Learnable State-Augmented Policies [7.512221808783587]
This paper addresses the challenge of packet-based information routing in large-scale wireless communication networks.<n>Opportunistic routing exploits the broadcast nature of wireless communication to dynamically select optimal forwarding nodes.<n>We propose a State-Augmentation (SA) based distributed optimization approach aimed at maximizing the total information handled by the source nodes in the network.
arXiv Detail & Related papers (2025-03-05T18:44:56Z) - Learning Load Balancing with GNN in MPTCP-Enabled Heterogeneous Networks [13.178956651532213]
We propose a graph neural network (GNN)-based model to tackle the LB problem for MP TCP-enabled HetNets.
Compared to the conventional deep neural network (DNN), the proposed GNN-based model exhibits two key strengths.
arXiv Detail & Related papers (2024-10-22T15:49:53Z) - Resource-Efficient Sensor Fusion via System-Wide Dynamic Gated Neural Networks [16.0018681576301]
We propose a novel algorithmic strategy called Quantile-constrained Inference (QIC)
QIC makes joint, high-quality, swift decisions on all the above aspects of the system.
Our results confirm that QIC matches the optimum and outperforms its alternatives by over 80%.
arXiv Detail & Related papers (2024-10-22T06:12:04Z) - SPARQ: Efficient Entanglement Distribution and Routing in Space-Air-Ground Quantum Networks [50.91365514137301]
Space-air-ground quantum (SPARQ) network is developed as a means for providing a seamless on-demand entanglement distribution.
Deep reinforcement learning framework is proposed and trained using deep Q-network (DQN) on multiple graphs of SPARQ.
Third-party entanglement distribution policy is proposed to establish entanglement between communication parties.
arXiv Detail & Related papers (2024-09-19T16:31:37Z) - SINR-Aware Deep Reinforcement Learning for Distributed Dynamic Channel
Allocation in Cognitive Interference Networks [10.514231683620517]
This paper focuses on real-world systems experiencing inter-carrier interference (ICI) and channel reuse by multiple large-scale networks.
We propose a novel multi-agent reinforcement learning framework for distributed DCA, named Channel Allocation RL To Overlapped Networks (CARLTON)
Our results demonstrate exceptional performance and robust generalization, showcasing superior efficiency compared to alternative state-of-the-art methods.
arXiv Detail & Related papers (2024-02-17T20:03:02Z) - Learning Resilient Radio Resource Management Policies with Graph Neural
Networks [124.89036526192268]
We formulate a resilient radio resource management problem with per-user minimum-capacity constraints.
We show that we can parameterize the user selection and power control policies using a finite set of parameters.
Thanks to such adaptation, our proposed method achieves a superior tradeoff between the average rate and the 5th percentile rate.
arXiv Detail & Related papers (2022-03-07T19:40:39Z) - Learning Autonomy in Management of Wireless Random Networks [102.02142856863563]
This paper presents a machine learning strategy that tackles a distributed optimization task in a wireless network with an arbitrary number of randomly interconnected nodes.
We develop a flexible deep neural network formalism termed distributed message-passing neural network (DMPNN) with forward and backward computations independent of the network topology.
arXiv Detail & Related papers (2021-06-15T09:03:28Z) - Resource Allocation via Graph Neural Networks in Free Space Optical
Fronthaul Networks [119.81868223344173]
This paper investigates the optimal resource allocation in free space optical (FSO) fronthaul networks.
We consider the graph neural network (GNN) for the policy parameterization to exploit the FSO network structure.
The primal-dual learning algorithm is developed to train the GNN in a model-free manner, where the knowledge of system models is not required.
arXiv Detail & Related papers (2020-06-26T14:20:48Z) - STDPG: A Spatio-Temporal Deterministic Policy Gradient Agent for Dynamic
Routing in SDN [6.27420060051673]
Dynamic routing in software-defined networking (SDN) can be viewed as a centralized decision-making problem.
We propose a novel model-free framework for dynamic routing in SDN, which is referred to as SDN-temporal deterministic policy gradient (STDPG) agent.
STDPG achieves better routing solutions in terms of average end-to-end delay.
arXiv Detail & Related papers (2020-04-21T07:19:07Z) - Deep Learning for Radio Resource Allocation with Diverse
Quality-of-Service Requirements in 5G [53.23237216769839]
We develop a deep learning framework to approximate the optimal resource allocation policy for base stations.
We find that a fully-connected neural network (NN) cannot fully guarantee the requirements due to the approximation errors and quantization errors of the numbers of subcarriers.
Considering that the distribution of wireless channels and the types of services in the wireless networks are non-stationary, we apply deep transfer learning to update NNs in non-stationary wireless networks.
arXiv Detail & Related papers (2020-03-29T04:48:22Z) - Toward fast and accurate human pose estimation via soft-gated skip
connections [97.06882200076096]
This paper is on highly accurate and highly efficient human pose estimation.
We re-analyze this design choice in the context of improving both the accuracy and the efficiency over the state-of-the-art.
Our model achieves state-of-the-art results on the MPII and LSP datasets.
arXiv Detail & Related papers (2020-02-25T18:51:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.