Distributed Uplink Beamforming in Cell-Free Networks Using Deep
Reinforcement Learning
- URL: http://arxiv.org/abs/2006.15138v2
- Date: Thu, 21 Oct 2021 20:32:14 GMT
- Title: Distributed Uplink Beamforming in Cell-Free Networks Using Deep
Reinforcement Learning
- Authors: Firas Fredj, Yasser Al-Eryani, Setareh Maghsudi, Mohamed Akrout, and
Ekram Hossain
- Abstract summary: We propose several beamforming techniques for an uplink cell-free network with centralized, semi-distributed, and fully distributed processing.
The proposed distributed beamforming technique performs better than the DDPG algorithm with centralized learning only for small-scale networks.
- Score: 25.579612460904873
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: The emergence of new wireless technologies together with the requirement of
massive connectivity results in several technical issues such as excessive
interference, high computational demand for signal processing, and lengthy
processing delays. In this work, we propose several beamforming techniques for
an uplink cell-free network with centralized, semi-distributed, and fully
distributed processing, all based on deep reinforcement learning (DRL). First,
we propose a fully centralized beamforming method that uses the deep
deterministic policy gradient algorithm (DDPG) with continuous space. We then
enhance this method by enabling distributed experience at access points (AP).
Indeed, we develop a beamforming scheme that uses the distributed
distributional deterministic policy gradients algorithm (D4PG) with the APs
representing the distributed agents. Finally, to decrease the computational
complexity, we propose a fully distributed beamforming scheme that divides the
beamforming computations among APs. The results show that the D4PG scheme with
distributed experience achieves the best performance irrespective of the
network size. Furthermore, the proposed distributed beamforming technique
performs better than the DDPG algorithm with centralized learning only for
small-scale networks. The performance superiority of the DDPG model becomes
more evident as the number of APs and/or users increases. Moreover, during the
operation stage, all DRL models demonstrate a significantly shorter processing
time than that of the conventional gradient descent (GD) solution.
Related papers
- FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - Unsupervised Graph-based Learning Method for Sub-band Allocation in 6G Subnetworks [2.0583251142940377]
We present an unsupervised approach for frequency sub-band allocation in wireless networks using graph-based learning.
We model the subnetwork deployment as a conflict graph and propose an unsupervised learning approach inspired by the graph colouring and the Potts model to optimize the sub-band allocation.
arXiv Detail & Related papers (2023-12-13T12:57:55Z) - Inter-Cell Network Slicing With Transfer Learning Empowered Multi-Agent
Deep Reinforcement Learning [6.523367518762879]
Network slicing enables operators to efficiently support diverse applications on a common physical infrastructure.
The ever-increasing densification of network deployment leads to complex and non-trivial inter-cell interference.
We develop a DIRP algorithm with multiple deep reinforcement learning (DRL) agents to cooperatively optimize resource partition in individual cells.
arXiv Detail & Related papers (2023-06-20T14:14:59Z) - Decentralized Federated Reinforcement Learning for User-Centric Dynamic
TFDD Control [37.54493447920386]
We propose a learning-based dynamic time-frequency division duplexing (D-TFDD) scheme to meet asymmetric and heterogeneous traffic demands.
We formulate the problem as a decentralized partially observable Markov decision process (Dec-POMDP)
In order to jointly optimize the global resources in a decentralized manner, we propose a federated reinforcement learning (RL) algorithm named Wolpertinger deep deterministic policy gradient (FWDDPG) algorithm.
arXiv Detail & Related papers (2022-11-04T07:39:21Z) - Adaptive Stochastic ADMM for Decentralized Reinforcement Learning in
Edge Industrial IoT [106.83952081124195]
Reinforcement learning (RL) has been widely investigated and shown to be a promising solution for decision-making and optimal control processes.
We propose an adaptive ADMM (asI-ADMM) algorithm and apply it to decentralized RL with edge-computing-empowered IIoT networks.
Experiment results show that our proposed algorithms outperform the state of the art in terms of communication costs and scalability, and can well adapt to complex IoT environments.
arXiv Detail & Related papers (2021-06-30T16:49:07Z) - Limited-Fronthaul Cell-Free Hybrid Beamforming with Distributed Deep
Neural Network [0.0]
Near-optimal solutions require a large amount of signaling exchange between access points (APs) and the network controller (NC)
We propose two unsupervised deep neural networks (DNN) architectures, fully and partially distributed, that can perform coordinated hybrid beamforming with zero or limited communication overhead between APs and NC.
arXiv Detail & Related papers (2021-06-30T16:42:32Z) - Adaptive Subcarrier, Parameter, and Power Allocation for Partitioned
Edge Learning Over Broadband Channels [69.18343801164741]
partitioned edge learning (PARTEL) implements parameter-server training, a well known distributed learning method, in wireless network.
We consider the case of deep neural network (DNN) models which can be trained using PARTEL by introducing some auxiliary variables.
arXiv Detail & Related papers (2020-10-08T15:27:50Z) - Optimization-driven Deep Reinforcement Learning for Robust Beamforming
in IRS-assisted Wireless Communications [54.610318402371185]
Intelligent reflecting surface (IRS) is a promising technology to assist downlink information transmissions from a multi-antenna access point (AP) to a receiver.
We minimize the AP's transmit power by a joint optimization of the AP's active beamforming and the IRS's passive beamforming.
We propose a deep reinforcement learning (DRL) approach that can adapt the beamforming strategies from past experiences.
arXiv Detail & Related papers (2020-05-25T01:42:55Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z) - Decentralized MCTS via Learned Teammate Models [89.24858306636816]
We present a trainable online decentralized planning algorithm based on decentralized Monte Carlo Tree Search.
We show that deep learning and convolutional neural networks can be employed to produce accurate policy approximators.
arXiv Detail & Related papers (2020-03-19T13:10:20Z) - Multiple Access in Dynamic Cell-Free Networks: Outage Performance and
Deep Reinforcement Learning-Based Design [24.632250413917816]
In future cell-free (or cell-less) wireless networks, a large number of devices in a geographical area will be served simultaneously by a large number of distributed access points (APs)
We propose a novel dynamic cell-free network architecture to reduce the complexity of joint processing of users' signals in presence of a large number of devices and APs.
In our system setting, the proposed DDPG-DDQN scheme is found to achieve around $78%$ of the rate achievable through an exhaustive search-based design.
arXiv Detail & Related papers (2020-01-29T03:00:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.