Path Design and Resource Management for NOMA enhanced Indoor Intelligent
Robots
- URL: http://arxiv.org/abs/2011.11745v2
- Date: Thu, 26 Nov 2020 23:25:46 GMT
- Title: Path Design and Resource Management for NOMA enhanced Indoor Intelligent
Robots
- Authors: Ruikang Zhong, Xiao Liu, Yuanwei Liu, Yue Chen and Xianbin Wang
- Abstract summary: A communication enabled indoor intelligent robots (IRs) service framework is proposed.
Lego modeling method is proposed, which can deterministically describe the indoor layout and channel state.
The investigated radio map is invoked as a virtual environment to train the reinforcement learning agent.
- Score: 58.980293789967575
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A communication enabled indoor intelligent robots (IRs) service framework is
proposed, where non-orthogonal multiple access (NOMA) technique is adopted to
enable highly reliable communications. In cooperation with the ultramodern
indoor channel model recently proposed by the International Telecommunication
Union (ITU), the Lego modeling method is proposed, which can deterministically
describe the indoor layout and channel state in order to construct the radio
map. The investigated radio map is invoked as a virtual environment to train
the reinforcement learning agent, which can save training time and hardware
costs. Build on the proposed communication model, motions of IRs who need to
reach designated mission destinations and their corresponding down-link power
allocation policy are jointly optimized to maximize the mission efficiency and
communication reliability of IRs. In an effort to solve this optimization
problem, a novel reinforcement learning approach named deep transfer
deterministic policy gradient (DT-DPG) algorithm is proposed. Our simulation
results demonstrate that 1) With the aid of NOMA techniques, the communication
reliability of IRs is effectively improved; 2) The radio map is qualified to be
a virtual training environment, and its statistical channel state information
improves training efficiency by about 30%; 3) The proposed DT-DPG algorithm is
superior to the conventional deep deterministic policy gradient (DDPG)
algorithm in terms of optimization performance, training time, and anti-local
optimum ability.
Related papers
- Enhancing Spectrum Efficiency in 6G Satellite Networks: A GAIL-Powered Policy Learning via Asynchronous Federated Inverse Reinforcement Learning [67.95280175998792]
A novel adversarial imitation learning (GAIL)-powered policy learning approach is proposed for optimizing beamforming, spectrum allocation, and remote user equipment (RUE) association ins.
We employ inverse RL (IRL) to automatically learn reward functions without manual tuning.
We show that the proposed MA-AL method outperforms traditional RL approaches, achieving a $14.6%$ improvement in convergence and reward value.
arXiv Detail & Related papers (2024-09-27T13:05:02Z) - Reconfigurable Intelligent Surface Aided Vehicular Edge Computing: Joint Phase-shift Optimization and Multi-User Power Allocation [28.47670676456068]
We introduce the use of Reconfigurable Intelligent Surfaces (RIS), which provide alternative communication pathways to assist vehicular communication.
We propose an innovative deep reinforcement learning (DRL) framework that combines the Deep Deterministic Policy Gradient (DDPG) algorithm for optimizing RIS phase-shift coefficients and the Multi-Agent Deep Deterministic Policy Gradient (MADDPG) algorithm for optimizing the power allocation of vehicle user (VU)
Simulation results show that our proposed scheme outperforms the traditional centralized DDPG, Twin Delayed Deep Deterministic Policy Gradient (TD3) and some typical schemes.
arXiv Detail & Related papers (2024-07-18T03:18:59Z) - Emergency Computing: An Adaptive Collaborative Inference Method Based on
Hierarchical Reinforcement Learning [14.929735103723573]
We propose an Emergency Network with Sensing, Communication, Computation, Caching, and Intelligence (E-SC3I)
The framework incorporates mechanisms for emergency computing, caching, integrated communication and sensing, and intelligence empowerment.
We specifically concentrate on emergency computing and propose an adaptive collaborative inference method (ACIM) based on hierarchical reinforcement learning.
arXiv Detail & Related papers (2024-02-03T13:28:35Z) - Muti-Agent Proximal Policy Optimization For Data Freshness in
UAV-assisted Networks [4.042622147977782]
We focus on the case where the collected data is time-sensitive, and it is critical to maintain its timeliness.
Our objective is to optimally design the UAVs' trajectories and the subsets of visited IoT devices such as the global Age-of-Updates (AoU) is minimized.
arXiv Detail & Related papers (2023-03-15T15:03:09Z) - Transferable Deep Reinforcement Learning Framework for Autonomous
Vehicles with Joint Radar-Data Communications [69.24726496448713]
We propose an intelligent optimization framework based on the Markov Decision Process (MDP) to help the AV make optimal decisions.
We then develop an effective learning algorithm leveraging recent advances of deep reinforcement learning techniques to find the optimal policy for the AV.
We show that the proposed transferable deep reinforcement learning framework reduces the obstacle miss detection probability by the AV up to 67% compared to other conventional deep reinforcement learning approaches.
arXiv Detail & Related papers (2021-05-28T08:45:37Z) - Reconfigurable Intelligent Surface Assisted Mobile Edge Computing with
Heterogeneous Learning Tasks [53.1636151439562]
Mobile edge computing (MEC) provides a natural platform for AI applications.
We present an infrastructure to perform machine learning tasks at an MEC with the assistance of a reconfigurable intelligent surface (RIS)
Specifically, we minimize the learning error of all participating users by jointly optimizing transmit power of mobile users, beamforming vectors of the base station, and the phase-shift matrix of the RIS.
arXiv Detail & Related papers (2020-12-25T07:08:50Z) - Distributional Reinforcement Learning for mmWave Communications with
Intelligent Reflectors on a UAV [119.97450366894718]
A novel communication framework that uses an unmanned aerial vehicle (UAV)-carried intelligent reflector (IR) is proposed.
In order to maximize the downlink sum-rate, the optimal precoding matrix (at the base station) and reflection coefficient (at the IR) are jointly derived.
arXiv Detail & Related papers (2020-11-03T16:50:37Z) - Optimization-driven Deep Reinforcement Learning for Robust Beamforming
in IRS-assisted Wireless Communications [54.610318402371185]
Intelligent reflecting surface (IRS) is a promising technology to assist downlink information transmissions from a multi-antenna access point (AP) to a receiver.
We minimize the AP's transmit power by a joint optimization of the AP's active beamforming and the IRS's passive beamforming.
We propose a deep reinforcement learning (DRL) approach that can adapt the beamforming strategies from past experiences.
arXiv Detail & Related papers (2020-05-25T01:42:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.