Distributed Reinforcement Learning for Flexible and Efficient UAV Swarm
Control
- URL: http://arxiv.org/abs/2103.04666v1
- Date: Mon, 8 Mar 2021 11:06:28 GMT
- Title: Distributed Reinforcement Learning for Flexible and Efficient UAV Swarm
Control
- Authors: Federico Venturini, Federico Mason, Francesco Pase, Federico
Chiariotti, Alberto Testolin, Andrea Zanella, Michele Zorzi
- Abstract summary: We propose a distributed Reinforcement Learning (RL) approach that scales to larger swarms without modifications.
Our experiments show that the proposed method can yield effective strategies, which are robust to communication channel impairments.
We also show that our approach achieves better performance compared to a computationally intensive look-ahead.
- Score: 28.463670610865837
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Over the past few years, the use of swarms of Unmanned Aerial Vehicles (UAVs)
in monitoring and remote area surveillance applications has become widespread
thanks to the price reduction and the increased capabilities of drones. The
drones in the swarm need to cooperatively explore an unknown area, in order to
identify and monitor interesting targets, while minimizing their movements. In
this work, we propose a distributed Reinforcement Learning (RL) approach that
scales to larger swarms without modifications. The proposed framework relies on
the possibility for the UAVs to exchange some information through a
communication channel, in order to achieve context-awareness and implicitly
coordinate the swarm's actions. Our experiments show that the proposed method
can yield effective strategies, which are robust to communication channel
impairments, and that can easily deal with non-uniform distributions of targets
and obstacles. Moreover, when agents are trained in a specific scenario, they
can adapt to a new one with minimal additional training. We also show that our
approach achieves better performance compared to a computationally intensive
look-ahead heuristic.
Related papers
- Predictive Covert Communication Against Multi-UAV Surveillance Using Graph Koopman Autoencoder [27.178522837149053]
Low Probability of Detection (LPD) communication aims to obscure the presence of radio frequency (RF) signals to evade surveillance.
accurately predicting future locations of UAVs is essential for enabling real-time LPD communication.
We introduce a novel framework termed predictive covert communication, aimed at minimizing detectability in terrestrial ad-hoc networks under multi-UAV surveillance.
arXiv Detail & Related papers (2024-09-25T16:02:45Z) - UAV-enabled Collaborative Beamforming via Multi-Agent Deep Reinforcement Learning [79.16150966434299]
We formulate a UAV-enabled collaborative beamforming multi-objective optimization problem (UCBMOP) to maximize the transmission rate of the UVAA and minimize the energy consumption of all UAVs.
We use the heterogeneous-agent trust region policy optimization (HATRPO) as the basic framework, and then propose an improved HATRPO algorithm, namely HATRPO-UCB.
arXiv Detail & Related papers (2024-04-11T03:19:22Z) - Anti-Jamming Path Planning Using GCN for Multi-UAV [0.0]
The effectiveness of UAV swarms can be severely compromised by jamming technology.
A novel approach, where UAV swarms leverage collective intelligence to predict jamming areas, is proposed.
A multi-agent control algorithm is then employed to disperse the UAV swarm, avoid jamming, and regroup upon reaching the target.
arXiv Detail & Related papers (2024-03-13T07:28:05Z) - Graph Koopman Autoencoder for Predictive Covert Communication Against
UAV Surveillance [29.15836826461713]
Low Probability of Detection (LPD) communication aims to obscure the very presence of radio frequency (RF) signals.
Unmanned Aerial Vehicles (UAVs) can detect RF signals from the ground by hovering over specific areas of interest.
We introduce a novel framework that combines graph neural networks (GNN) with Koopman theory to predict the trajectories of multiple fixed-wing UAVs.
arXiv Detail & Related papers (2024-01-23T23:42:55Z) - Aerial View Goal Localization with Reinforcement Learning [6.165163123577484]
We present a framework that emulates a search-and-rescue (SAR)-like setup without requiring access to actual UAVs.
In this framework, an agent operates on top of an aerial image (proxy for a search area) and is tasked with localizing a goal that is described in terms of visual cues.
We propose AiRLoc, a reinforcement learning (RL)-based model that decouples exploration (searching for distant goals) and exploitation (localizing nearby goals)
arXiv Detail & Related papers (2022-09-08T10:27:53Z) - Collaborative Target Search with a Visual Drone Swarm: An Adaptive
Curriculum Embedded Multistage Reinforcement Learning Approach [0.0]
We propose a novel data-efficient deep reinforcement learning (DRL) approach called adaptive curriculum embedded multistage learning (ACEMSL)
We decompose the collaborative target search task into several subtasks including individual obstacle avoidance, target search, and inter-agent collaboration, and progressively train the agents with multistage learning.
We deploy the trained model over a real visual drone swarm and perform CTS operations without fine-tuning.
arXiv Detail & Related papers (2022-04-26T09:32:22Z) - 3D UAV Trajectory and Data Collection Optimisation via Deep
Reinforcement Learning [75.78929539923749]
Unmanned aerial vehicles (UAVs) are now beginning to be deployed for enhancing the network performance and coverage in wireless communication.
It is challenging to obtain an optimal resource allocation scheme for the UAV-assisted Internet of Things (IoT)
In this paper, we design a new UAV-assisted IoT systems relying on the shortest flight path of the UAVs while maximising the amount of data collected from IoT devices.
arXiv Detail & Related papers (2021-06-06T14:08:41Z) - Transferable Deep Reinforcement Learning Framework for Autonomous
Vehicles with Joint Radar-Data Communications [69.24726496448713]
We propose an intelligent optimization framework based on the Markov Decision Process (MDP) to help the AV make optimal decisions.
We then develop an effective learning algorithm leveraging recent advances of deep reinforcement learning techniques to find the optimal policy for the AV.
We show that the proposed transferable deep reinforcement learning framework reduces the obstacle miss detection probability by the AV up to 67% compared to other conventional deep reinforcement learning approaches.
arXiv Detail & Related papers (2021-05-28T08:45:37Z) - Efficient UAV Trajectory-Planning using Economic Reinforcement Learning [65.91405908268662]
We introduce REPlanner, a novel reinforcement learning algorithm inspired by economic transactions to distribute tasks between UAVs.
We formulate the path planning problem as a multi-agent economic game, where agents can cooperate and compete for resources.
As the system computes task distributions via UAV cooperation, it is highly resilient to any change in the swarm size.
arXiv Detail & Related papers (2021-03-03T20:54:19Z) - Adversarial Attacks On Multi-Agent Communication [80.4392160849506]
Modern autonomous systems will soon be deployed at scale, opening up the possibility for cooperative multi-agent systems.
Such advantages rely heavily on communication channels which have been shown to be vulnerable to security breaches.
In this paper, we explore such adversarial attacks in a novel multi-agent setting where agents communicate by sharing learned intermediate representations.
arXiv Detail & Related papers (2021-01-17T00:35:26Z) - Learning to Communicate and Correct Pose Errors [75.03747122616605]
We study the setting proposed in V2VNet, where nearby self-driving vehicles jointly perform object detection and motion forecasting in a cooperative manner.
We propose a novel neural reasoning framework that learns to communicate, to estimate potential errors, and to reach a consensus about those errors.
arXiv Detail & Related papers (2020-11-10T18:19:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.