Deep Reinforcement Learning in mmW-NOMA: Joint Power Allocation and
Hybrid Beamforming
- URL: http://arxiv.org/abs/2205.06814v1
- Date: Fri, 13 May 2022 07:55:48 GMT
- Title: Deep Reinforcement Learning in mmW-NOMA: Joint Power Allocation and
Hybrid Beamforming
- Authors: Abbas Akbarpour-Kasgari, Mehrdad Ardebilipour
- Abstract summary: High demand of data rate could be ensured by Non-Orthogonal Multiple Access (NOMA) approach in the millimetre-wave (mmW) frequency band.
Joint power allocation and hybrid beamforming of mmW-NOMA systems is brought up via recent advances in machine learning and control theory approaches.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: High demand of data rate in the next generation of wireless communication
could be ensured by Non-Orthogonal Multiple Access (NOMA) approach in the
millimetre-wave (mmW) frequency band. Decreasing the interference on the other
users while maintaining the bit rate via joint power allocation and beamforming
is mandatory to guarantee the high demand of bit-rate. Furthermore, mmW
frequency bands dictates the hybrid structure for beamforming because of the
trade-off in implementation and performance, simultaneously. In this paper,
joint power allocation and hybrid beamforming of mmW-NOMA systems is brought up
via recent advances in machine learning and control theory approaches called
Deep Reinforcement Learning (DRL). Actor-critic phenomena is exploited to
measure the immediate reward and providing the new action to maximize the
overall Q-value of the network. Additionally, to improve the stability of the
approach, we have utilized Soft Actor-Critic (SAC) approach where overall
reward and action entropy is maximized, simultaneously. The immediate reward
has been defined based on the soft weighted summation of the rate of all the
users. The soft weighting is based on the achieved rate and allocated power of
each user. Furthermore, the channel responses between the users and base
station (BS) is defined as the state of environment, while action space is
involved of the digital and analog beamforming weights and allocated power to
each user. The simulation results represent the superiority of the proposed
approach rather than the Time-Division Multiple Access (TDMA) and Non-Line of
Sight (NLOS)-NOMA in terms of sum-rate of the users. It's outperformance is
caused by the joint optimization and independency of the proposed approach to
the channel responses.
Related papers
- High Efficiency Inference Accelerating Algorithm for NOMA-based Mobile
Edge Computing [23.88527790721402]
Splitting the inference model between device, edge server, and cloud can improve the performance of EI greatly.
NOMA, which is the key supporting technologies of B5G/6G, can achieve massive connections and high spectrum efficiency.
We propose the effective communication and computing resource allocation algorithm to accelerate the model inference at edge.
arXiv Detail & Related papers (2023-12-26T02:05:52Z) - Wirelessly Powered Federated Learning Networks: Joint Power Transfer,
Data Sensing, Model Training, and Resource Allocation [24.077525032187893]
Federated learning (FL) has found many successes in wireless networks.
implementation of FL has been hindered by the energy limitation of mobile devices (MDs) and the availability of training data at MDs.
How to integrate wireless power transfer and sustainable sustainable FL networks.
arXiv Detail & Related papers (2023-08-09T13:38:58Z) - Active RIS-aided EH-NOMA Networks: A Deep Reinforcement Learning
Approach [66.53364438507208]
An active reconfigurable intelligent surface (RIS)-aided multi-user downlink communication system is investigated.
Non-orthogonal multiple access (NOMA) is employed to improve spectral efficiency, and the active RIS is powered by energy harvesting (EH)
An advanced LSTM based algorithm is developed to predict users' dynamic communication state.
A DDPG based algorithm is proposed to joint control the amplification matrix and phase shift matrix RIS.
arXiv Detail & Related papers (2023-04-11T13:16:28Z) - Joint Power Allocation and Beamformer for mmW-NOMA Downlink Systems by
Deep Reinforcement Learning [0.0]
Joint power allocation and beamforming of mmW-NOMA systems is mandatory.
We have exploited Deep Reinforcement Learning (DRL) approach due to policy generation leading to an optimized sum-rate of users.
arXiv Detail & Related papers (2022-05-13T07:42:03Z) - Federated Learning for Energy-limited Wireless Networks: A Partial Model
Aggregation Approach [79.59560136273917]
limited communication resources, bandwidth and energy, and data heterogeneity across devices are main bottlenecks for federated learning (FL)
We first devise a novel FL framework with partial model aggregation (PMA)
The proposed PMA-FL improves 2.72% and 11.6% accuracy on two typical heterogeneous datasets.
arXiv Detail & Related papers (2022-04-20T19:09:52Z) - Collaborative Intelligent Reflecting Surface Networks with Multi-Agent
Reinforcement Learning [63.83425382922157]
Intelligent reflecting surface (IRS) is envisioned to be widely applied in future wireless networks.
In this paper, we investigate a multi-user communication system assisted by cooperative IRS devices with the capability of energy harvesting.
arXiv Detail & Related papers (2022-03-26T20:37:14Z) - Deep Reinforcement Learning Based Multidimensional Resource Management
for Energy Harvesting Cognitive NOMA Communications [64.1076645382049]
Combination of energy harvesting (EH), cognitive radio (CR), and non-orthogonal multiple access (NOMA) is a promising solution to improve energy efficiency.
In this paper, we study the spectrum, energy, and time resource management for deterministic-CR-NOMA IoT systems.
arXiv Detail & Related papers (2021-09-17T08:55:48Z) - Optimal Power Allocation for Rate Splitting Communications with Deep
Reinforcement Learning [61.91604046990993]
This letter introduces a novel framework to optimize the power allocation for users in a Rate Splitting Multiple Access network.
In the network, messages intended for users are split into different parts that are a single common part and respective private parts.
arXiv Detail & Related papers (2021-07-01T06:32:49Z) - Low-Latency Federated Learning over Wireless Channels with Differential
Privacy [142.5983499872664]
In federated learning (FL), model training is distributed over clients and local models are aggregated by a central server.
In this paper, we aim to minimize FL training delay over wireless channels, constrained by overall training performance as well as each client's differential privacy (DP) requirement.
arXiv Detail & Related papers (2021-06-20T13:51:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.