Performance Optimization of Energy-Harvesting Underlay Cognitive Radio Networks Using Reinforcement Learning
- URL: http://arxiv.org/abs/2505.14581v1
- Date: Tue, 20 May 2025 16:38:32 GMT
- Title: Performance Optimization of Energy-Harvesting Underlay Cognitive Radio Networks Using Reinforcement Learning
- Authors: Deemah H. Tashman, Soumaya Cherkaoui, Walaa Hamouda,
- Abstract summary: A reinforcement learning technique is employed to maximize the performance of a cognitive radio network.<n>Our approach outperforms a baseline strategy and converges, as shown by our findings.
- Score: 16.57893415196489
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, a reinforcement learning technique is employed to maximize the performance of a cognitive radio network (CRN). In the presence of primary users (PUs), it is presumed that two secondary users (SUs) access the licensed band within underlay mode. In addition, the SU transmitter is assumed to be an energy-constrained device that requires harvesting energy in order to transmit signals to their intended destination. Therefore, we propose that there are two main sources of energy; the interference of PUs' transmissions and ambient radio frequency (RF) sources. The SU will select whether to gather energy from PUs or only from ambient sources based on a predetermined threshold. The process of energy harvesting from the PUs' messages is accomplished via the time switching approach. In addition, based on a deep Q-network (DQN) approach, the SU transmitter determines whether to collect energy or transmit messages during each time slot as well as selects the suitable transmission power in order to maximize its average data rate. Our approach outperforms a baseline strategy and converges, as shown by our findings.
Related papers
- Neuromorphic Split Computing with Wake-Up Radios: Architecture and Design via Digital Twinning [97.99077847606624]
This work proposes a novel architecture that integrates a wake-up radio mechanism within a split computing system consisting of remote, wirelessly connected, NPUs.
A key challenge in the design of a wake-up radio-based neuromorphic split computing system is the selection of thresholds for sensing, wake-up signal detection, and decision making.
arXiv Detail & Related papers (2024-04-02T10:19:04Z) - Multiagent Reinforcement Learning with an Attention Mechanism for
Improving Energy Efficiency in LoRa Networks [52.96907334080273]
As the network scale increases, the energy efficiency of LoRa networks decreases sharply due to severe packet collisions.
We propose a transmission parameter allocation algorithm based on multiagent reinforcement learning (MALoRa)
Simulation results demonstrate that MALoRa significantly improves the system EE compared with baseline algorithms.
arXiv Detail & Related papers (2023-09-16T11:37:23Z) - Sum-Rate Maximization of RSMA-based Aerial Communications with Energy
Harvesting: A Reinforcement Learning Approach [5.35414932422173]
A self-sustainable aerial base station serves multiple users by utilizing the harvested energy.
Considering maximizing the sum-rate from the long-term perspective, we utilize a deep reinforcement learning (DRL) approach.
We show the superiority of the proposed scheme over several baseline methods in terms of the average sum-rate performance.
arXiv Detail & Related papers (2023-06-22T15:38:22Z) - Deep Reinforcement Learning Based Multidimensional Resource Management
for Energy Harvesting Cognitive NOMA Communications [64.1076645382049]
Combination of energy harvesting (EH), cognitive radio (CR), and non-orthogonal multiple access (NOMA) is a promising solution to improve energy efficiency.
In this paper, we study the spectrum, energy, and time resource management for deterministic-CR-NOMA IoT systems.
arXiv Detail & Related papers (2021-09-17T08:55:48Z) - Optimal Power Allocation for Rate Splitting Communications with Deep
Reinforcement Learning [61.91604046990993]
This letter introduces a novel framework to optimize the power allocation for users in a Rate Splitting Multiple Access network.
In the network, messages intended for users are split into different parts that are a single common part and respective private parts.
arXiv Detail & Related papers (2021-07-01T06:32:49Z) - Wirelessly Powered Federated Edge Learning: Optimal Tradeoffs Between
Convergence and Power Transfer [42.30741737568212]
We propose the solution of powering devices using wireless power transfer (WPT)
This work aims at the derivation of guidelines on deploying the resultant wirelessly powered FEEL (WP-FEEL) system.
The results provide useful guidelines on WPT provisioning to provide a guaranteer on learning performance.
arXiv Detail & Related papers (2021-02-24T15:47:34Z) - Learning Centric Power Allocation for Edge Intelligence [84.16832516799289]
Edge intelligence has been proposed, which collects distributed data and performs machine learning at the edge.
This paper proposes a learning centric power allocation (LCPA) method, which allocates radio resources based on an empirical classification error model.
Experimental results show that the proposed LCPA algorithm significantly outperforms other power allocation algorithms.
arXiv Detail & Related papers (2020-07-21T07:02:07Z) - Cognitive Radio Network Throughput Maximization with Deep Reinforcement
Learning [58.44609538048923]
Radio Frequency powered Cognitive Radio Networks (RF-CRN) are likely to be the eyes and ears of upcoming modern networks such as Internet of Things (IoT)
To be considered autonomous, the RF-powered network entities need to make decisions locally to maximize the network throughput under the uncertainty of any network environment.
In this paper, deep reinforcement learning is proposed to overcome the shortcomings and allow a wireless gateway to derive an optimal policy to maximize network throughput.
arXiv Detail & Related papers (2020-07-07T01:49:07Z) - Learning to Charge RF-Energy Harvesting Devices in WiFi Networks [0.0]
We propose two solutions that enable the AP to manage its harvested energy via transmit power control.
The first solution uses a deep Q-network (DQN) whilst the second solution uses Model Predictive Control (MPC) to control the AP's transmit power.
Our results show that our DQN and MPC solutions improve energy efficiency and user satisfaction by 16% to 35%, and 10% to 42% as compared to competing algorithms.
arXiv Detail & Related papers (2020-05-25T10:55:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.