Detection and Tracking Meet Drones Challenge
- URL: http://arxiv.org/abs/2001.06303v3
- Date: Mon, 4 Oct 2021 03:37:37 GMT
- Title: Detection and Tracking Meet Drones Challenge
- Authors: Pengfei Zhu, Longyin Wen, Dawei Du, Xiao Bian, Heng Fan, Qinghua Hu,
Haibin Ling
- Abstract summary: This paper presents a review of object detection and tracking datasets and benchmarks, and discusses the challenges of collecting large-scale drone-based object detection and tracking datasets with manual annotations.
We describe our VisDrone dataset, which is captured over various urban/suburban areas of 14 different cities across China from North to South.
We provide a detailed analysis of the current state of the field of large-scale object detection and tracking on drones, and conclude the challenge as well as propose future directions.
- Score: 131.31749447313197
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Drones, or general UAVs, equipped with cameras have been fast deployed with a
wide range of applications, including agriculture, aerial photography, and
surveillance. Consequently, automatic understanding of visual data collected
from drones becomes highly demanding, bringing computer vision and drones more
and more closely. To promote and track the developments of object detection and
tracking algorithms, we have organized three challenge workshops in conjunction
with ECCV 2018, ICCV 2019 and ECCV 2020, attracting more than 100 teams around
the world. We provide a large-scale drone captured dataset, VisDrone, which
includes four tracks, i.e., (1) image object detection, (2) video object
detection, (3) single object tracking, and (4) multi-object tracking. In this
paper, we first present a thorough review of object detection and tracking
datasets and benchmarks, and discuss the challenges of collecting large-scale
drone-based object detection and tracking datasets with fully manual
annotations. After that, we describe our VisDrone dataset, which is captured
over various urban/suburban areas of 14 different cities across China from
North to South. Being the largest such dataset ever published, VisDrone enables
extensive evaluation and investigation of visual analysis algorithms for the
drone platform. We provide a detailed analysis of the current state of the
field of large-scale object detection and tracking on drones, and conclude the
challenge as well as propose future directions. We expect the benchmark largely
boost the research and development in video analysis on drone platforms. All
the datasets and experimental results can be downloaded from
https://github.com/VisDrone/VisDrone-Dataset.
Related papers
- C2FDrone: Coarse-to-Fine Drone-to-Drone Detection using Vision Transformer Networks [23.133250476580038]
A vision-based drone-to-drone detection system is crucial for various applications like collision avoidance, countering hostile drones, and search-and-rescue operations.
detecting drones presents unique challenges, including small object sizes, distortion, and real-time processing requirements.
We propose a novel coarse-to-fine detection strategy based on vision transformers.
arXiv Detail & Related papers (2024-04-30T05:51:21Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - TransVisDrone: Spatio-Temporal Transformer for Vision-based
Drone-to-Drone Detection in Aerial Videos [57.92385818430939]
Drone-to-drone detection using visual feed has crucial applications, such as detecting drone collisions, detecting drone attacks, or coordinating flight with other drones.
Existing methods are computationally costly, follow non-end-to-end optimization, and have complex multi-stage pipelines, making them less suitable for real-time deployment on edge devices.
We propose a simple yet effective framework, itTransVisDrone, that provides an end-to-end solution with higher computational efficiency.
arXiv Detail & Related papers (2022-10-16T03:05:13Z) - GIAOTracker: A comprehensive framework for MCMOT with global information
and optimizing strategies in VisDrone 2021 [6.4515884598231]
We propose a new multiple object tracker, named GIAOTracker.
It consists of three stages, i.e., online tracking, global link and post-processing.
With the effectiveness of the three stages, GIAOTracker achieves state-of-the-art performance on the VisDrone MOT dataset.
arXiv Detail & Related papers (2022-02-24T09:42:00Z) - DronePose: The identification, segmentation, and orientation detection
of drones via neural networks [3.161871054978445]
We present a CNN using a decision tree and ensemble structure to fully characterise drones in flight.
Our system determines the drone type, orientation (in terms of pitch, roll, and yaw), and performs segmentation to classify different body parts.
We also provide a computer model for the rapid generation of large quantities of accurately labelled photo-realistic training data.
arXiv Detail & Related papers (2021-12-10T12:34:53Z) - A dataset for multi-sensor drone detection [67.75999072448555]
The use of small and remotely controlled unmanned aerial vehicles (UAVs) has increased in recent years.
Most studies on drone detection fail to specify the type of acquisition device, the drone type, the detection range, or the dataset.
We contribute with an annotated multi-sensor database for drone detection that includes infrared and visible videos and audio files.
arXiv Detail & Related papers (2021-11-02T20:52:03Z) - Scarce Data Driven Deep Learning of Drones via Generalized Data
Distribution Space [12.377024173799631]
We show how understanding the general distribution of the drone data via a Generative Adversarial Network (GAN) can allow us to acquire missing data to achieve rapid and more accurate learning.
We demonstrate our results on a drone image dataset, which contains both real drone images as well as simulated images from computer-aided design.
arXiv Detail & Related papers (2021-08-18T17:07:32Z) - Dogfight: Detecting Drones from Drones Videos [58.158988162743825]
This paper attempts to address the problem of drones detection from other flying drones variations.
The erratic movement of the source and target drones, small size, arbitrary shape, large intensity, and occlusion make this problem quite challenging.
To handle this, instead of using region-proposal based methods, we propose to use a two-stage segmentation-based approach.
arXiv Detail & Related papers (2021-03-31T17:43:31Z) - University-1652: A Multi-view Multi-source Benchmark for Drone-based
Geo-localization [87.74121935246937]
We introduce a new multi-view benchmark for drone-based geo-localization, named University-1652.
University-1652 contains data from three platforms, i.e., synthetic drones, satellites and ground cameras of 1,652 university buildings around the world.
Experiments show that University-1652 helps the model to learn the viewpoint-invariant features and also has good generalization ability in the real-world scenario.
arXiv Detail & Related papers (2020-02-27T15:24:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.