Q-Learning based system for path planning with unmanned aerial vehicles
swarms in obstacle environments
- URL: http://arxiv.org/abs/2303.17655v2
- Date: Fri, 25 Aug 2023 13:42:26 GMT
- Title: Q-Learning based system for path planning with unmanned aerial vehicles
swarms in obstacle environments
- Authors: Alejandro Puente-Castro, Daniel Rivero, Eurico Pedrosa, Artur Pereira,
Nuno Lau, Enrique Fernandez-Blanco
- Abstract summary: A Reinforcement Learning based system is proposed for solving this problem in environments with obstacles by making use of Q-Learning.
The goal of these paths is to ensure complete coverage of an area with fixed obstacles for tasks, like field prospecting.
The results are satisfactory, showing that the system obtains solutions in fewer movements the more UAVs there are.
- Score: 38.82157836789187
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Path Planning methods for autonomous control of Unmanned Aerial Vehicle (UAV)
swarms are on the rise because of all the advantages they bring. There are more
and more scenarios where autonomous control of multiple UAVs is required. Most
of these scenarios present a large number of obstacles, such as power lines or
trees. If all UAVs can be operated autonomously, personnel expenses can be
decreased. In addition, if their flight paths are optimal, energy consumption
is reduced. This ensures that more battery time is left for other operations.
In this paper, a Reinforcement Learning based system is proposed for solving
this problem in environments with obstacles by making use of Q-Learning. This
method allows a model, in this particular case an Artificial Neural Network, to
self-adjust by learning from its mistakes and achievements. Regardless of the
size of the map or the number of UAVs in the swarm, the goal of these paths is
to ensure complete coverage of an area with fixed obstacles for tasks, like
field prospecting. Setting goals or having any prior information aside from the
provided map is not required. For experimentation, five maps of different sizes
with different obstacles were used. The experiments were performed with
different number of UAVs. For the calculation of the results, the number of
actions taken by all UAVs to complete the task in each experiment is taken into
account. The lower the number of actions, the shorter the path and the lower
the energy consumption. The results are satisfactory, showing that the system
obtains solutions in fewer movements the more UAVs there are. For a better
presentation, these results have been compared to another state-of-the-art
approach.
Related papers
- Autonomous Decision Making for UAV Cooperative Pursuit-Evasion Game with Reinforcement Learning [50.33447711072726]
This paper proposes a deep reinforcement learning-based model for decision-making in multi-role UAV cooperative pursuit-evasion game.
The proposed method enables autonomous decision-making of the UAVs in pursuit-evasion game scenarios.
arXiv Detail & Related papers (2024-11-05T10:45:30Z) - Tiny Multi-Agent DRL for Twins Migration in UAV Metaverses: A Multi-Leader Multi-Follower Stackelberg Game Approach [57.15309977293297]
The synergy between Unmanned Aerial Vehicles (UAVs) and metaverses is giving rise to an emerging paradigm named UAV metaverses.
We propose a tiny machine learning-based Stackelberg game framework based on pruning techniques for efficient UT migration in UAV metaverses.
arXiv Detail & Related papers (2024-01-18T02:14:13Z) - UAV Obstacle Avoidance by Human-in-the-Loop Reinforcement in Arbitrary
3D Environment [17.531224704021273]
This paper focuses on the continuous control of the unmanned aerial vehicle (UAV) based on a deep reinforcement learning method.
We propose a deep reinforcement learning (DRL)-based method combined with human-in-the-loop, which allows the UAV to avoid obstacles automatically during flying.
arXiv Detail & Related papers (2023-04-07T01:44:05Z) - Solving reward-collecting problems with UAVs: a comparison of online
optimization and Q-learning [2.4251007104039006]
We study the problem of identifying a short path from a designated start to a goal, while collecting all rewards and avoiding adversaries that move randomly on the grid.
We present a comparison of three methods to solve this problem: namely we implement a Deep Q-Learning model, an $varepsilon$-greedy tabular Q-Learning model, and an online optimization framework.
Our experiments, designed using simple grid-world environments with random adversaries, showcase how these approaches work and compare them in terms of performance, accuracy, and computational time.
arXiv Detail & Related papers (2021-11-30T22:27:24Z) - Advanced Algorithms of Collision Free Navigation and Flocking for
Autonomous UAVs [0.0]
This report contributes towards the state-of-the-art in UAV control for safe autonomous navigation and motion coordination of multi-UAV systems.
The first part of this report deals with single-UAV systems. The complex problem of three-dimensional (3D) collision-free navigation in unknown/dynamic environments is addressed.
The second part of this report addresses safe navigation for multi-UAV systems. Distributed motion coordination methods of multi-UAV systems for flocking and 3D area coverage are developed.
arXiv Detail & Related papers (2021-10-30T03:51:40Z) - A Multi-UAV System for Exploration and Target Finding in Cluttered and
GPS-Denied Environments [68.31522961125589]
We propose a framework for a team of UAVs to cooperatively explore and find a target in complex GPS-denied environments with obstacles.
The team of UAVs autonomously navigates, explores, detects, and finds the target in a cluttered environment with a known map.
Results indicate that the proposed multi-UAV system has improvements in terms of time-cost, the proportion of search area surveyed, as well as successful rates for search and rescue missions.
arXiv Detail & Related papers (2021-07-19T12:54:04Z) - 3D UAV Trajectory and Data Collection Optimisation via Deep
Reinforcement Learning [75.78929539923749]
Unmanned aerial vehicles (UAVs) are now beginning to be deployed for enhancing the network performance and coverage in wireless communication.
It is challenging to obtain an optimal resource allocation scheme for the UAV-assisted Internet of Things (IoT)
In this paper, we design a new UAV-assisted IoT systems relying on the shortest flight path of the UAVs while maximising the amount of data collected from IoT devices.
arXiv Detail & Related papers (2021-06-06T14:08:41Z) - Efficient UAV Trajectory-Planning using Economic Reinforcement Learning [65.91405908268662]
We introduce REPlanner, a novel reinforcement learning algorithm inspired by economic transactions to distribute tasks between UAVs.
We formulate the path planning problem as a multi-agent economic game, where agents can cooperate and compete for resources.
As the system computes task distributions via UAV cooperation, it is highly resilient to any change in the swarm size.
arXiv Detail & Related papers (2021-03-03T20:54:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.