Muti-Agent Proximal Policy Optimization For Data Freshness in
UAV-assisted Networks
- URL: http://arxiv.org/abs/2303.08680v1
- Date: Wed, 15 Mar 2023 15:03:09 GMT
- Title: Muti-Agent Proximal Policy Optimization For Data Freshness in
UAV-assisted Networks
- Authors: Mouhamed Naby Ndiaye, El Houcine Bergou, Hajar El Hammouti
- Abstract summary: We focus on the case where the collected data is time-sensitive, and it is critical to maintain its timeliness.
Our objective is to optimally design the UAVs' trajectories and the subsets of visited IoT devices such as the global Age-of-Updates (AoU) is minimized.
- Score: 4.042622147977782
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unmanned aerial vehicles (UAVs) are seen as a promising technology to perform
a wide range of tasks in wireless communication networks. In this work, we
consider the deployment of a group of UAVs to collect the data generated by IoT
devices. Specifically, we focus on the case where the collected data is
time-sensitive, and it is critical to maintain its timeliness. Our objective is
to optimally design the UAVs' trajectories and the subsets of visited IoT
devices such as the global Age-of-Updates (AoU) is minimized. To this end, we
formulate the studied problem as a mixed-integer nonlinear programming (MINLP)
under time and quality of service constraints. To efficiently solve the
resulting optimization problem, we investigate the cooperative Multi-Agent
Reinforcement Learning (MARL) framework and propose an RL approach based on the
popular on-policy Reinforcement Learning (RL) algorithm: Policy Proximal
Optimization (PPO). Our approach leverages the centralized training
decentralized execution (CTDE) framework where the UAVs learn their optimal
policies while training a centralized value function. Our simulation results
show that the proposed MAPPO approach reduces the global AoU by at least a
factor of 1/2 compared to conventional off-policy reinforcement learning
approaches.
Related papers
- Joint Demonstration and Preference Learning Improves Policy Alignment with Human Feedback [58.049113055986375]
We develop a single stage approach named Alignment with Integrated Human Feedback (AIHF) to train reward models and the policy.
The proposed approach admits a suite of efficient algorithms, which can easily reduce to, and leverage, popular alignment algorithms.
We demonstrate the efficiency of the proposed solutions with extensive experiments involving alignment problems in LLMs and robotic control problems in MuJoCo.
arXiv Detail & Related papers (2024-06-11T01:20:53Z) - UAV-enabled Collaborative Beamforming via Multi-Agent Deep Reinforcement Learning [79.16150966434299]
We formulate a UAV-enabled collaborative beamforming multi-objective optimization problem (UCBMOP) to maximize the transmission rate of the UVAA and minimize the energy consumption of all UAVs.
We use the heterogeneous-agent trust region policy optimization (HATRPO) as the basic framework, and then propose an improved HATRPO algorithm, namely HATRPO-UCB.
arXiv Detail & Related papers (2024-04-11T03:19:22Z) - Multi-Agent Reinforcement Learning for Offloading Cellular Communications with Cooperating UAVs [21.195346908715972]
Unmanned aerial vehicles present an alternative means to offload data traffic from terrestrial BSs.
This paper presents a novel approach to efficiently serve multiple UAVs for data offloading from terrestrial BSs.
arXiv Detail & Related papers (2024-02-05T12:36:08Z) - Personalized Federated Deep Reinforcement Learning-based Trajectory
Optimization for Multi-UAV Assisted Edge Computing [22.09756306579992]
UAVs can serve as intelligent servers in edge computing environments, optimizing their flight trajectories to maximize communication system throughput.
Deep reinforcement learning (DRL)-based trajectory optimization algorithms may suffer from poor training performance due to intricate terrain features and inadequate training data.
This work proposes a novel solution, namely personalized federated deep reinforcement learning (PF-DRL), for multi-UAV trajectory optimization.
arXiv Detail & Related papers (2023-09-05T12:54:40Z) - Multi-Objective Optimization for UAV Swarm-Assisted IoT with Virtual
Antenna Arrays [55.736718475856726]
Unmanned aerial vehicle (UAV) network is a promising technology for assisting Internet-of-Things (IoT)
Existing UAV-assisted data harvesting and dissemination schemes require UAVs to frequently fly between the IoTs and access points.
We introduce collaborative beamforming into IoTs and UAVs simultaneously to achieve energy and time-efficient data harvesting and dissemination.
arXiv Detail & Related papers (2023-08-03T02:49:50Z) - AI-based Radio and Computing Resource Allocation and Path Planning in
NOMA NTNs: AoI Minimization under CSI Uncertainty [23.29963717212139]
We develop a hierarchical aerial computing framework composed of high altitude platform (HAP) and unmanned aerial vehicles (UAVs)
It is shown that task scheduling significantly reduces the average AoI.
It is shown that power allocation has a marginal effect on the average AoI compared to using full transmission power for all users.
arXiv Detail & Related papers (2023-05-01T11:52:15Z) - DL-DRL: A double-level deep reinforcement learning approach for
large-scale task scheduling of multi-UAV [65.07776277630228]
We propose a double-level deep reinforcement learning (DL-DRL) approach based on a divide and conquer framework (DCF)
Particularly, we design an encoder-decoder structured policy network in our upper-level DRL model to allocate the tasks to different UAVs.
We also exploit another attention based policy network in our lower-level DRL model to construct the route for each UAV, with the objective to maximize the number of executed tasks.
arXiv Detail & Related papers (2022-08-04T04:35:53Z) - Optimization for Master-UAV-powered Auxiliary-Aerial-IRS-assisted IoT
Networks: An Option-based Multi-agent Hierarchical Deep Reinforcement
Learning Approach [56.84948632954274]
This paper investigates a master unmanned aerial vehicle (MUAV)-powered Internet of Things (IoT) network.
We propose using a rechargeable auxiliary UAV (AUAV) equipped with an intelligent reflecting surface (IRS) to enhance the communication signals from the MUAV.
Under the proposed model, we investigate the optimal collaboration strategy of these energy-limited UAVs to maximize the accumulated throughput of the IoT network.
arXiv Detail & Related papers (2021-12-20T15:45:28Z) - 3D UAV Trajectory and Data Collection Optimisation via Deep
Reinforcement Learning [75.78929539923749]
Unmanned aerial vehicles (UAVs) are now beginning to be deployed for enhancing the network performance and coverage in wireless communication.
It is challenging to obtain an optimal resource allocation scheme for the UAV-assisted Internet of Things (IoT)
In this paper, we design a new UAV-assisted IoT systems relying on the shortest flight path of the UAVs while maximising the amount of data collected from IoT devices.
arXiv Detail & Related papers (2021-06-06T14:08:41Z) - Multi-UAV Path Planning for Wireless Data Harvesting with Deep
Reinforcement Learning [18.266087952180733]
We propose a multi-agent reinforcement learning (MARL) approach that can adapt to profound changes in the scenario parameters defining the data harvesting mission.
We show that our proposed network architecture enables the agents to cooperate effectively by carefully dividing the data collection task among themselves.
arXiv Detail & Related papers (2020-10-23T14:59:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.