Aerial View Goal Localization with Reinforcement Learning
- URL: http://arxiv.org/abs/2209.03694v1
- Date: Thu, 8 Sep 2022 10:27:53 GMT
- Title: Aerial View Goal Localization with Reinforcement Learning
- Authors: Aleksis Pirinen, Anton Samuelsson, John Backsund, Kalle {\AA}str\"om
- Abstract summary: We present a framework that emulates a search-and-rescue (SAR)-like setup without requiring access to actual UAVs.
In this framework, an agent operates on top of an aerial image (proxy for a search area) and is tasked with localizing a goal that is described in terms of visual cues.
We propose AiRLoc, a reinforcement learning (RL)-based model that decouples exploration (searching for distant goals) and exploitation (localizing nearby goals)
- Score: 6.165163123577484
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With an increased amount and availability of unmanned aerial vehicles (UAVs)
and other remote sensing devices (e.g. satellites), we have recently seen a
vast increase in computer vision methods for aerial view data. One application
of such technologies is within search-and-rescue (SAR), where the task is to
localize and assist one or several people who are missing, for example after a
natural disaster. In many cases the rough location may be known and a UAV can
be deployed to explore a given, confined area to precisely localize the missing
people. Due to time and battery constraints it is often critical that
localization is performed as efficiently as possible. In this work, we approach
this type of problem by abstracting it as an aerial view goal localization task
in a framework that emulates a SAR-like setup without requiring access to
actual UAVs. In this framework, an agent operates on top of an aerial image
(proxy for a search area) and is tasked with localizing a goal that is
described in terms of visual cues. To further mimic the situation on an actual
UAV, the agent is not able to observe the search area in its entirety, not even
at low resolution, and thus it has to operate solely based on partial glimpses
when navigating towards the goal. To tackle this task, we propose AiRLoc, a
reinforcement learning (RL)-based model that decouples exploration (searching
for distant goals) and exploitation (localizing nearby goals). Extensive
evaluations show that AiRLoc outperforms heuristic search methods as well as
alternative learnable approaches. We also conduct a proof-of-concept study
which indicates that the learnable methods outperform humans on average. Code
has been made publicly available: https://github.com/aleksispi/airloc.
Related papers
- GOMAA-Geo: GOal Modality Agnostic Active Geo-localization [49.599465495973654]
We consider the task of active geo-localization (AGL) in which an agent uses a sequence of visual cues observed during aerial navigation to find a target specified through multiple possible modalities.
GOMAA-Geo is a goal modality active geo-localization agent for zero-shot generalization between different goal modalities.
arXiv Detail & Related papers (2024-06-04T02:59:36Z) - Joint Path planning and Power Allocation of a Cellular-Connected UAV
using Apprenticeship Learning via Deep Inverse Reinforcement Learning [7.760962597460447]
This paper investigates an interference-aware joint path planning and power allocation mechanism for a cellular-connected unmanned aerial vehicle (UAV) in a sparse suburban environment.
The UAV aims to maximize its uplink throughput and minimize the level of interference to the ground user equipment (UEs) connected to the neighbor cellular BSs.
An apprenticeship learning method is utilized via inverse reinforcement learning (IRL) based on both Q-learning and deep reinforcement learning (DRL)
arXiv Detail & Related papers (2023-06-15T20:50:05Z) - Rethinking Drone-Based Search and Rescue with Aerial Person Detection [79.76669658740902]
The visual inspection of aerial drone footage is an integral part of land search and rescue (SAR) operations today.
We propose a novel deep learning algorithm to automate this aerial person detection (APD) task.
We present the novel Aerial Inspection RetinaNet (AIR) algorithm as the combination of these contributions.
arXiv Detail & Related papers (2021-11-17T21:48:31Z) - A Multi-UAV System for Exploration and Target Finding in Cluttered and
GPS-Denied Environments [68.31522961125589]
We propose a framework for a team of UAVs to cooperatively explore and find a target in complex GPS-denied environments with obstacles.
The team of UAVs autonomously navigates, explores, detects, and finds the target in a cluttered environment with a known map.
Results indicate that the proposed multi-UAV system has improvements in terms of time-cost, the proportion of search area surveyed, as well as successful rates for search and rescue missions.
arXiv Detail & Related papers (2021-07-19T12:54:04Z) - Distributed Reinforcement Learning for Flexible and Efficient UAV Swarm
Control [28.463670610865837]
We propose a distributed Reinforcement Learning (RL) approach that scales to larger swarms without modifications.
Our experiments show that the proposed method can yield effective strategies, which are robust to communication channel impairments.
We also show that our approach achieves better performance compared to a computationally intensive look-ahead.
arXiv Detail & Related papers (2021-03-08T11:06:28Z) - Batch Exploration with Examples for Scalable Robotic Reinforcement
Learning [63.552788688544254]
Batch Exploration with Examples (BEE) explores relevant regions of the state-space guided by a modest number of human provided images of important states.
BEE is able to tackle challenging vision-based manipulation tasks both in simulation and on a real Franka robot.
arXiv Detail & Related papers (2020-10-22T17:49:25Z) - POMP: Pomcp-based Online Motion Planning for active visual search in
indoor environments [89.43830036483901]
We focus on the problem of learning an optimal policy for Active Visual Search (AVS) of objects in known indoor environments with an online setup.
Our POMP method uses as input the current pose of an agent and a RGB-D frame.
We validate our method on the publicly available AVD benchmark, achieving an average success rate of 0.76 with an average path length of 17.1.
arXiv Detail & Related papers (2020-09-17T08:23:50Z) - Perceiving Traffic from Aerial Images [86.994032967469]
We propose an object detection method called Butterfly Detector that is tailored to detect objects in aerial images.
We evaluate our Butterfly Detector on two publicly available UAV datasets (UAVDT and VisDrone 2019) and show that it outperforms previous state-of-the-art methods while remaining real-time.
arXiv Detail & Related papers (2020-09-16T11:37:43Z) - Reinforcement Learning for UAV Autonomous Navigation, Mapping and Target
Detection [36.79380276028116]
We study a joint detection, mapping and navigation problem for a single unmanned aerial vehicle (UAV) equipped with a low complexity radar and flying in an unknown environment.
The goal is to optimize its trajectory with the purpose of maximizing the mapping accuracy and to avoid areas where measurements might not be sufficiently informative from the perspective of a target detection.
arXiv Detail & Related papers (2020-05-05T20:39:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.