Reinforcement Learning for Agile Active Target Sensing with a UAV
- URL: http://arxiv.org/abs/2212.08214v1
- Date: Fri, 16 Dec 2022 01:01:17 GMT
- Title: Reinforcement Learning for Agile Active Target Sensing with a UAV
- Authors: Harsh Goel, Laura Jarin Lipschitz, Saurav Agarwal, Sandeep Manjanna,
and Vijay Kumar
- Abstract summary: This paper develops a deep reinforcement learning approach to plan informative trajectories.
It exploits its current belief of the target states and incorporates inaccurate sensor models for high-fidelity classification.
A unique characteristic of our approach is that it is robust to varying amounts of deviations from the true target distribution.
- Score: 10.070339628481445
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Active target sensing is the task of discovering and classifying an unknown
number of targets in an environment and is critical in search-and-rescue
missions. This paper develops a deep reinforcement learning approach to plan
informative trajectories that increase the likelihood for an uncrewed aerial
vehicle (UAV) to discover missing targets. Our approach efficiently (1)
explores the environment to discover new targets, (2) exploits its current
belief of the target states and incorporates inaccurate sensor models for
high-fidelity classification, and (3) generates dynamically feasible
trajectories for an agile UAV by employing a motion primitive library.
Extensive simulations on randomly generated environments show that our approach
is more efficient in discovering and classifying targets than several other
baselines. A unique characteristic of our approach, in contrast to heuristic
informative path planning approaches, is that it is robust to varying amounts
of deviations of the prior belief from the true target distribution, thereby
alleviating the challenge of designing heuristics specific to the application
conditions.
Related papers
- FIT-SLAM -- Fisher Information and Traversability estimation-based
Active SLAM for exploration in 3D environments [1.4474137122906163]
Active visual SLAM finds a wide array of applications in-Denied sub-terrain environments and outdoor environments for ground robots.
It is imperative to incorporate the perception considerations in the goal selection and path planning towards the goal during an exploration mission.
We propose FIT-SLAM, a new exploration method tailored for unmanned ground vehicles (UGVs) to explore 3D environments.
arXiv Detail & Related papers (2024-01-17T16:46:38Z) - Enhancing Infrared Small Target Detection Robustness with Bi-Level
Adversarial Framework [61.34862133870934]
We propose a bi-level adversarial framework to promote the robustness of detection in the presence of distinct corruptions.
Our scheme remarkably improves 21.96% IOU across a wide array of corruptions and notably promotes 4.97% IOU on the general benchmark.
arXiv Detail & Related papers (2023-09-03T06:35:07Z) - Distributed multi-agent target search and tracking with Gaussian process
and reinforcement learning [26.499110405106812]
We propose a multi-agent reinforcement learning technique with target map building based on distributed process.
We evaluate the performance and transferability of the trained policy in simulation and demonstrate the method on a swarm of micro unmanned aerial vehicles.
arXiv Detail & Related papers (2023-08-29T01:53:14Z) - Discrete Factorial Representations as an Abstraction for Goal
Conditioned Reinforcement Learning [99.38163119531745]
We show that applying a discretizing bottleneck can improve performance in goal-conditioned RL setups.
We experimentally prove the expected return on out-of-distribution goals, while still allowing for specifying goals with expressive structure.
arXiv Detail & Related papers (2022-11-01T03:31:43Z) - Uncertainty with UAV Search of Multiple Goal-oriented Targets [25.918290198644122]
This paper considers the complex problem of a team of UAVs searching targets under uncertainty.
We suggest a real-time algorithmic framework for the UAVs, combining entropy andtemporal belief.
We have empirically evaluated the algorithmic framework, and have shown its efficiency and significant performance improvement.
arXiv Detail & Related papers (2022-03-03T09:57:00Z) - Generative multitask learning mitigates target-causing confounding [61.21582323566118]
We propose a simple and scalable approach to causal representation learning for multitask learning.
The improvement comes from mitigating unobserved confounders that cause the targets, but not the input.
Our results on the Attributes of People and Taskonomy datasets reflect the conceptual improvement in robustness to prior probability shift.
arXiv Detail & Related papers (2022-02-08T20:42:14Z) - Goal-Aware Cross-Entropy for Multi-Target Reinforcement Learning [15.33496710690063]
We propose goal-aware cross-entropy (GACE) loss, that can be utilized in a self-supervised way.
We then devise goal-discriminative attention networks (GDAN) which utilize the goal-relevant information to focus on the given instruction.
arXiv Detail & Related papers (2021-10-25T14:24:39Z) - Adversarial Intrinsic Motivation for Reinforcement Learning [60.322878138199364]
We investigate whether the Wasserstein-1 distance between a policy's state visitation distribution and a target distribution can be utilized effectively for reinforcement learning tasks.
Our approach, termed Adversarial Intrinsic Motivation (AIM), estimates this Wasserstein-1 distance through its dual objective and uses it to compute a supplemental reward function.
arXiv Detail & Related papers (2021-05-27T17:51:34Z) - Learning to Track Dynamic Targets in Partially Known Environments [48.49957897251128]
We use a deep reinforcement learning approach to solve active target tracking.
In particular, we introduce Active Tracking Target Network (ATTN), a unified RL policy that is capable of solving major sub-tasks of active target tracking.
arXiv Detail & Related papers (2020-06-17T22:45:24Z) - Automatic Curriculum Learning through Value Disagreement [95.19299356298876]
Continually solving new, unsolved tasks is the key to learning diverse behaviors.
In the multi-task domain, where an agent needs to reach multiple goals, the choice of training goals can largely affect sample efficiency.
We propose setting up an automatic curriculum for goals that the agent needs to solve.
We evaluate our method across 13 multi-goal robotic tasks and 5 navigation tasks, and demonstrate performance gains over current state-of-the-art methods.
arXiv Detail & Related papers (2020-06-17T03:58:25Z) - Reinforcement Learning for UAV Autonomous Navigation, Mapping and Target
Detection [36.79380276028116]
We study a joint detection, mapping and navigation problem for a single unmanned aerial vehicle (UAV) equipped with a low complexity radar and flying in an unknown environment.
The goal is to optimize its trajectory with the purpose of maximizing the mapping accuracy and to avoid areas where measurements might not be sufficiently informative from the perspective of a target detection.
arXiv Detail & Related papers (2020-05-05T20:39:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.