Collaborative Target Search with a Visual Drone Swarm: An Adaptive
Curriculum Embedded Multistage Reinforcement Learning Approach
- URL: http://arxiv.org/abs/2204.12181v3
- Date: Sat, 25 Nov 2023 07:11:25 GMT
- Title: Collaborative Target Search with a Visual Drone Swarm: An Adaptive
Curriculum Embedded Multistage Reinforcement Learning Approach
- Authors: Jiaping Xiao, Phumrapee Pisutsin and Mir Feroskhan
- Abstract summary: We propose a novel data-efficient deep reinforcement learning (DRL) approach called adaptive curriculum embedded multistage learning (ACEMSL)
We decompose the collaborative target search task into several subtasks including individual obstacle avoidance, target search, and inter-agent collaboration, and progressively train the agents with multistage learning.
We deploy the trained model over a real visual drone swarm and perform CTS operations without fine-tuning.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Equipping drones with target search capabilities is highly desirable for
applications in disaster rescue and smart warehouse delivery systems. Multiple
intelligent drones that can collaborate with each other and maneuver among
obstacles show more effectiveness in accomplishing tasks in a shorter amount of
time. However, carrying out collaborative target search (CTS) without prior
target information is extremely challenging, especially with a visual drone
swarm. In this work, we propose a novel data-efficient deep reinforcement
learning (DRL) approach called adaptive curriculum embedded multistage learning
(ACEMSL) to address these challenges, mainly 3-D sparse reward space
exploration with limited visual perception and collaborative behavior
requirements. Specifically, we decompose the CTS task into several subtasks
including individual obstacle avoidance, target search, and inter-agent
collaboration, and progressively train the agents with multistage learning.
Meanwhile, an adaptive embedded curriculum (AEC) is designed, where the task
difficulty level (TDL) can be adaptively adjusted based on the success rate
(SR) achieved in training. ACEMSL allows data-efficient training and
individual-team reward allocation for the visual drone swarm. Furthermore, we
deploy the trained model over a real visual drone swarm and perform CTS
operations without fine-tuning. Extensive simulations and real-world flight
tests validate the effectiveness and generalizability of ACEMSL. The project is
available at https://github.com/NTU-UAVG/CTS-visual-drone-swarm.git.
Related papers
- Energy-Aware Multi-Agent Reinforcement Learning for Collaborative Execution in Mission-Oriented Drone Networks [3.4918110778972458]
Mission-oriented drone networks have been widely used for structural inspection, disaster monitoring, border surveillance, etc.
Due to the limited battery capacity of drones, mission execution strategy impacts network performance and mission completion.
We leverage multi-agent reinforcement learning (MARL) to manage the challenge in this study.
arXiv Detail & Related papers (2024-10-29T22:43:26Z) - A Dual Curriculum Learning Framework for Multi-UAV Pursuit-Evasion in Diverse Environments [15.959963737956848]
This paper addresses multi-UAV pursuit-evasion, where a group of drones cooperate to capture a fast evader in a confined environment with obstacles.
Existing algorithms, which simplify the pursuit-evasion problem, often lack expressive coordination strategies and struggle to capture the evader in extreme scenarios.
We introduce a dual curriculum learning framework, named DualCL, which addresses multi-UAV pursuit-evasion in diverse environments and demonstrates zero-shot transfer ability to unseen scenarios.
arXiv Detail & Related papers (2023-12-19T15:39:09Z) - Efficient Adaptive Human-Object Interaction Detection with
Concept-guided Memory [64.11870454160614]
We propose an efficient Adaptive HOI Detector with Concept-guided Memory (ADA-CM)
ADA-CM has two operating modes. The first mode makes it tunable without learning new parameters in a training-free paradigm.
Our proposed method achieves competitive results with state-of-the-art on the HICO-DET and V-COCO datasets with much less training time.
arXiv Detail & Related papers (2023-09-07T13:10:06Z) - Multi-Objective Optimization for UAV Swarm-Assisted IoT with Virtual
Antenna Arrays [55.736718475856726]
Unmanned aerial vehicle (UAV) network is a promising technology for assisting Internet-of-Things (IoT)
Existing UAV-assisted data harvesting and dissemination schemes require UAVs to frequently fly between the IoTs and access points.
We introduce collaborative beamforming into IoTs and UAVs simultaneously to achieve energy and time-efficient data harvesting and dissemination.
arXiv Detail & Related papers (2023-08-03T02:49:50Z) - Learning Multi-Pursuit Evasion for Safe Targeted Navigation of Drones [0.0]
This paper proposes a novel approach, asynchronous multi-stage deep reinforcement learning (AMS-DRL), to train adversarial neural networks.
AMS-DRL evolves adversarial agents in a pursuit-evasion game where the pursuers and the evader are asynchronously trained in a bipartite graph way.
We evaluate our method in extensive simulations and show that it outperforms baselines with higher navigation success rates.
arXiv Detail & Related papers (2023-04-07T01:59:16Z) - Effective Adaptation in Multi-Task Co-Training for Unified Autonomous
Driving [103.745551954983]
In this paper, we investigate the transfer performance of various types of self-supervised methods, including MoCo and SimCLR, on three downstream tasks.
We find that their performances are sub-optimal or even lag far behind the single-task baseline.
We propose a simple yet effective pretrain-adapt-finetune paradigm for general multi-task training.
arXiv Detail & Related papers (2022-09-19T12:15:31Z) - DL-DRL: A double-level deep reinforcement learning approach for
large-scale task scheduling of multi-UAV [65.07776277630228]
We propose a double-level deep reinforcement learning (DL-DRL) approach based on a divide and conquer framework (DCF)
Particularly, we design an encoder-decoder structured policy network in our upper-level DRL model to allocate the tasks to different UAVs.
We also exploit another attention based policy network in our lower-level DRL model to construct the route for each UAV, with the objective to maximize the number of executed tasks.
arXiv Detail & Related papers (2022-08-04T04:35:53Z) - Decentralized Reinforcement Learning for Multi-Target Search and
Detection by a Team of Drones [12.055303570215335]
Targets search and detection encompasses a variety of decision problems such as coverage, surveillance, search, observing and pursuit-evasion.
We develop a multi-agent deep reinforcement learning (MADRL) method to coordinate a group of aerial vehicles (drones) for the purpose of locating a set of static targets in an unknown area.
arXiv Detail & Related papers (2021-03-17T09:04:47Z) - Distributed Reinforcement Learning for Flexible and Efficient UAV Swarm
Control [28.463670610865837]
We propose a distributed Reinforcement Learning (RL) approach that scales to larger swarms without modifications.
Our experiments show that the proposed method can yield effective strategies, which are robust to communication channel impairments.
We also show that our approach achieves better performance compared to a computationally intensive look-ahead.
arXiv Detail & Related papers (2021-03-08T11:06:28Z) - Batch Exploration with Examples for Scalable Robotic Reinforcement
Learning [63.552788688544254]
Batch Exploration with Examples (BEE) explores relevant regions of the state-space guided by a modest number of human provided images of important states.
BEE is able to tackle challenging vision-based manipulation tasks both in simulation and on a real Franka robot.
arXiv Detail & Related papers (2020-10-22T17:49:25Z) - Planning to Explore via Self-Supervised World Models [120.31359262226758]
Plan2Explore is a self-supervised reinforcement learning agent.
We present a new approach to self-supervised exploration and fast adaptation to new tasks.
Without any training supervision or task-specific interaction, Plan2Explore outperforms prior self-supervised exploration methods.
arXiv Detail & Related papers (2020-05-12T17:59:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.