Towards Multi-agent Reinforcement Learning for Wireless Network Protocol
Synthesis
- URL: http://arxiv.org/abs/2102.01611v1
- Date: Tue, 2 Feb 2021 17:13:37 GMT
- Title: Towards Multi-agent Reinforcement Learning for Wireless Network Protocol
Synthesis
- Authors: Hrishikesh Dutta and Subir Biswas
- Abstract summary: This paper proposes a multi-agent reinforcement learning based medium access framework for wireless networks.
The access problem is formulated as a Markov Decision Process (MDP), and solved using reinforcement learning with every network node acting as a distributed learning agent.
- Score: 2.6397379133308214
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper proposes a multi-agent reinforcement learning based medium access
framework for wireless networks. The access problem is formulated as a Markov
Decision Process (MDP), and solved using reinforcement learning with every
network node acting as a distributed learning agent. The solution components
are developed step by step, starting from a single-node access scenario in
which a node agent incrementally learns to control MAC layer packet loads for
reining in self-collisions. The strategy is then scaled up for multi-node
fully-connected scenarios by using more elaborate reward structures. It also
demonstrates preliminary feasibility for more general partially connected
topologies. It is shown that by learning to adjust MAC layer transmission
probabilities, the protocol is not only able to attain theoretical maximum
throughput at an optimal load, but unlike classical approaches, it can also
retain that maximum throughput at higher loading conditions. Additionally, the
mechanism is agnostic to heterogeneous loading while preserving that feature.
It is also shown that access priorities of the protocol across nodes can be
parametrically adjusted. Finally, it is also shown that the online learning
feature of reinforcement learning is able to make the protocol adapt to
time-varying loading conditions.
Related papers
- Joint Admission Control and Resource Allocation of Virtual Network Embedding via Hierarchical Deep Reinforcement Learning [69.00997996453842]
We propose a deep Reinforcement Learning approach to learn a joint Admission Control and Resource Allocation policy for virtual network embedding.
We show that HRL-ACRA outperforms state-of-the-art baselines in terms of both the acceptance ratio and long-term average revenue.
arXiv Detail & Related papers (2024-06-25T07:42:30Z) - Design Optimization of NOMA Aided Multi-STAR-RIS for Indoor Environments: A Convex Approximation Imitated Reinforcement Learning Approach [51.63921041249406]
Non-orthogonal multiple access (NOMA) enables multiple users to share the same frequency band, and simultaneously transmitting and reflecting reconfigurable intelligent surface (STAR-RIS)
deploying STAR-RIS indoors presents challenges in interference mitigation, power consumption, and real-time configuration.
A novel network architecture utilizing multiple access points (APs), STAR-RISs, and NOMA is proposed for indoor communication.
arXiv Detail & Related papers (2024-06-19T07:17:04Z) - Sparsity-Aware Intelligent Massive Random Access Control in Open RAN: A
Reinforcement Learning Based Approach [61.74489383629319]
Massive random access of devices in the emerging Open Radio Access Network (O-RAN) brings great challenge to the access control and management.
reinforcement-learning (RL)-assisted scheme of closed-loop access control is proposed to preserve sparsity of access requests.
Deep-RL-assisted SAUD is proposed to resolve highly complex environments with continuous and high-dimensional state and action spaces.
arXiv Detail & Related papers (2023-03-05T12:25:49Z) - Reinforcement Learning for Protocol Synthesis in Resource-Constrained
Wireless Sensor and IoT Networks [1.462434043267217]
The paper introduces the use of RL and Multi Armed Bandit (MAB), a specific type of RL, for Medium Access Control (MAC)
It then introduces a novel learning based protocol synthesis framework that addresses specific difficulties and limitations in medium access for both random access and time slotted networks.
The ability of independent protocol learning by the nodes makes the system robust and adaptive to the changes in network and traffic conditions.
arXiv Detail & Related papers (2023-01-14T03:28:26Z) - Artificial Intelligence Empowered Multiple Access for Ultra Reliable and
Low Latency THz Wireless Networks [76.89730672544216]
Terahertz (THz) wireless networks are expected to catalyze the beyond fifth generation (B5G) era.
To satisfy the ultra-reliability and low-latency demands of several B5G applications, novel mobility management approaches are required.
This article presents a holistic MAC layer approach that enables intelligent user association and resource allocation, as well as flexible and adaptive mobility management.
arXiv Detail & Related papers (2022-08-17T03:00:24Z) - Online Distributed Evolutionary Optimization of Time Division Multiple
Access Protocols [4.87717454493713]
We envision a protocol as an emergent property of a network, obtained by an environment-driven Distributed Hill Climbing algorithm.
We show how Distributed Hill Climbing can reach different trade-offs in terms of energy consumption and protocol performance.
arXiv Detail & Related papers (2022-04-27T20:47:48Z) - Cooperative Multi-Agent Reinforcement Learning Based Distributed Dynamic
Spectrum Access in Cognitive Radio Networks [46.723006378363785]
Dynamic spectrum access (DSA) is a promising paradigm to remedy the problem of inefficient spectrum utilization.
In this paper, we investigate the distributed DSA problem for multi-user in a typical cognitive radio network.
We employ the deep recurrent Q-network (DRQN) to address the partial observability of the state for each cognitive user.
arXiv Detail & Related papers (2021-06-17T06:52:21Z) - Medium Access using Distributed Reinforcement Learning for IoTs with
Low-Complexity Wireless Transceivers [2.6397379133308214]
This paper proposes a distributed Reinforcement Learning (RL) based framework that can be used for MAC layer wireless protocols in IoT networks with low-complexity wireless transceivers.
In this framework, the access protocols are first formulated as Markov Decision Processes (MDP) and then solved using RL.
The paper demonstrates the performance of the learning paradigm and its abilities to make nodes adapt their optimal transmission strategies on the fly in response to various network dynamics.
arXiv Detail & Related papers (2021-04-29T17:57:43Z) - All at Once Network Quantization via Collaborative Knowledge Transfer [56.95849086170461]
We develop a novel collaborative knowledge transfer approach for efficiently training the all-at-once quantization network.
Specifically, we propose an adaptive selection strategy to choose a high-precision enquoteteacher for transferring knowledge to the low-precision student.
To effectively transfer knowledge, we develop a dynamic block swapping method by randomly replacing the blocks in the lower-precision student network with the corresponding blocks in the higher-precision teacher network.
arXiv Detail & Related papers (2021-03-02T03:09:03Z) - Scalable Deep Reinforcement Learning for Routing and Spectrum Access in
Physical Layer [12.018165291620164]
We propose a novel reinforcement learning approach for simultaneous routing and spectrum access in wireless ad-hoc networks.
A single agent makes all routing and spectrum access decisions as it moves along the frontier nodes of each flow.
The proposed deep reinforcement learning strategy is capable of accounting for the mutual interference between the links.
arXiv Detail & Related papers (2020-12-22T01:47:20Z) - Deep Multi-Task Learning for Cooperative NOMA: System Design and
Principles [52.79089414630366]
We develop a novel deep cooperative NOMA scheme, drawing upon the recent advances in deep learning (DL)
We develop a novel hybrid-cascaded deep neural network (DNN) architecture such that the entire system can be optimized in a holistic manner.
arXiv Detail & Related papers (2020-07-27T12:38:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.