Multi-Drone based Single Object Tracking with Agent Sharing Network
- URL: http://arxiv.org/abs/2003.06994v1
- Date: Mon, 16 Mar 2020 03:27:04 GMT
- Title: Multi-Drone based Single Object Tracking with Agent Sharing Network
- Authors: Pengfei Zhu, Jiayu Zheng, Dawei Du, Longyin Wen, Yiming Sun, Qinghua
Hu
- Abstract summary: Multi-Drone single Object Tracking dataset consists of 92 groups of video clips with 113,918 high resolution frames taken by two drones and 63 groups of video clips with 145,875 high resolution frames taken by three drones.
Agent sharing network (ASNet) is proposed by self-supervised template sharing and view-aware fusion of the target from multiple drones.
- Score: 74.8198920355117
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Drone equipped with cameras can dynamically track the target in the air from
a broader view compared with static cameras or moving sensors over the ground.
However, it is still challenging to accurately track the target using a single
drone due to several factors such as appearance variations and severe
occlusions. In this paper, we collect a new Multi-Drone single Object Tracking
(MDOT) dataset that consists of 92 groups of video clips with 113,918 high
resolution frames taken by two drones and 63 groups of video clips with 145,875
high resolution frames taken by three drones. Besides, two evaluation metrics
are specially designed for multi-drone single object tracking, i.e. automatic
fusion score (AFS) and ideal fusion score (IFS). Moreover, an agent sharing
network (ASNet) is proposed by self-supervised template sharing and view-aware
fusion of the target from multiple drones, which can improve the tracking
accuracy significantly compared with single drone tracking. Extensive
experiments on MDOT show that our ASNet significantly outperforms recent
state-of-the-art trackers.
Related papers
- RockTrack: A 3D Robust Multi-Camera-Ken Multi-Object Tracking Framework [28.359633046753228]
We propose RockTrack, a 3D MOT method for multi-camera detectors.
RockTrack incorporates a confidence-guided preprocessing module to extract reliable motion and image observations.
RockTrack achieves state-of-the-art performance on the nuScenes vision-only tracking leaderboard with 59.1% AMOTA.
arXiv Detail & Related papers (2024-09-18T07:08:08Z) - DroneMOT: Drone-based Multi-Object Tracking Considering Detection Difficulties and Simultaneous Moving of Drones and Objects [6.449663756698312]
Multi-object tracking (MOT) on static platforms, such as surveillance cameras, has achieved significant progress.
However, the effectiveness of traditional MOT methods is significantly reduced when it comes to dynamic platforms like drones.
This paper proposes DroneMOT, which considers the fast movements of drones to enhance the drone-based object detection and feature embedding for small-sized, blurred, and occluded objects.
arXiv Detail & Related papers (2024-07-12T07:18:18Z) - C2FDrone: Coarse-to-Fine Drone-to-Drone Detection using Vision Transformer Networks [23.133250476580038]
A vision-based drone-to-drone detection system is crucial for various applications like collision avoidance, countering hostile drones, and search-and-rescue operations.
detecting drones presents unique challenges, including small object sizes, distortion, and real-time processing requirements.
We propose a novel coarse-to-fine detection strategy based on vision transformers.
arXiv Detail & Related papers (2024-04-30T05:51:21Z) - TransVisDrone: Spatio-Temporal Transformer for Vision-based
Drone-to-Drone Detection in Aerial Videos [57.92385818430939]
Drone-to-drone detection using visual feed has crucial applications, such as detecting drone collisions, detecting drone attacks, or coordinating flight with other drones.
Existing methods are computationally costly, follow non-end-to-end optimization, and have complex multi-stage pipelines, making them less suitable for real-time deployment on edge devices.
We propose a simple yet effective framework, itTransVisDrone, that provides an end-to-end solution with higher computational efficiency.
arXiv Detail & Related papers (2022-10-16T03:05:13Z) - EVPropNet: Detecting Drones By Finding Propellers For Mid-Air Landing
And Following [11.79762223888294]
Drone propellers are the fastest moving parts of an image and cannot be directly "seen" by a classical camera without severe motion blur.
We train a deep neural network called EVPropNet to detect propellers from the data of an event camera.
We present two applications of our network: (a) tracking and following an unmarked drone and (b) landing on a near-hover drone.
arXiv Detail & Related papers (2021-06-29T01:16:01Z) - Detection, Tracking, and Counting Meets Drones in Crowds: A Benchmark [97.07865343576361]
We construct a benchmark with a new drone-captured largescale dataset, named as DroneCrowd.
We annotate 20,800 people trajectories with 4.8 million heads and several video-level attributes.
We design the Space-Time Neighbor-Aware Network (STNNet) as a strong baseline to solve object detection, tracking and counting jointly in dense crowds.
arXiv Detail & Related papers (2021-05-06T04:46:14Z) - Tracking-by-Counting: Using Network Flows on Crowd Density Maps for
Tracking Multiple Targets [96.98888948518815]
State-of-the-art multi-object tracking(MOT) methods follow the tracking-by-detection paradigm.
We propose a new MOT paradigm, tracking-by-counting, tailored for crowded scenes.
arXiv Detail & Related papers (2020-07-18T19:51:53Z) - DroTrack: High-speed Drone-based Object Tracking Under Uncertainty [0.23204178451683263]
DroTrack is a high-speed visual single-object tracking framework for drone-captured video sequences.
We implement an effective object segmentation based on Fuzzy C Means.
We also leverage the geometrical angular motion to estimate a reliable object scale.
arXiv Detail & Related papers (2020-05-02T13:16:16Z) - University-1652: A Multi-view Multi-source Benchmark for Drone-based
Geo-localization [87.74121935246937]
We introduce a new multi-view benchmark for drone-based geo-localization, named University-1652.
University-1652 contains data from three platforms, i.e., synthetic drones, satellites and ground cameras of 1,652 university buildings around the world.
Experiments show that University-1652 helps the model to learn the viewpoint-invariant features and also has good generalization ability in the real-world scenario.
arXiv Detail & Related papers (2020-02-27T15:24:15Z) - Detection and Tracking Meet Drones Challenge [131.31749447313197]
This paper presents a review of object detection and tracking datasets and benchmarks, and discusses the challenges of collecting large-scale drone-based object detection and tracking datasets with manual annotations.
We describe our VisDrone dataset, which is captured over various urban/suburban areas of 14 different cities across China from North to South.
We provide a detailed analysis of the current state of the field of large-scale object detection and tracking on drones, and conclude the challenge as well as propose future directions.
arXiv Detail & Related papers (2020-01-16T00:11:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.