Archangel: A Hybrid UAV-based Human Detection Benchmark with Position
and Pose Metadata
- URL: http://arxiv.org/abs/2209.00128v3
- Date: Tue, 8 Aug 2023 18:48:21 GMT
- Title: Archangel: A Hybrid UAV-based Human Detection Benchmark with Position
and Pose Metadata
- Authors: Yi-Ting Shen, Yaesop Lee, Heesung Kwon, Damon M. Conover, Shuvra S.
Bhattacharyya, Nikolas Vale, Joshua D. Gray, G. Jeremy Leong, Kenneth
Evensen, Frank Skirlo
- Abstract summary: Archangel is the first UAV-based object detection dataset composed of real and synthetic subsets.
A series of experiments are carefully designed with a state-of-the-art object detector to demonstrate the benefits of leveraging the metadata.
- Score: 10.426019628829204
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning to detect objects, such as humans, in imagery captured by an
unmanned aerial vehicle (UAV) usually suffers from tremendous variations caused
by the UAV's position towards the objects. In addition, existing UAV-based
benchmark datasets do not provide adequate dataset metadata, which is essential
for precise model diagnosis and learning features invariant to those
variations. In this paper, we introduce Archangel, the first UAV-based object
detection dataset composed of real and synthetic subsets captured with similar
imagining conditions and UAV position and object pose metadata. A series of
experiments are carefully designed with a state-of-the-art object detector to
demonstrate the benefits of leveraging the metadata during model evaluation.
Moreover, several crucial insights involving both real and synthetic data
during model optimization are presented. In the end, we discuss the advantages,
limitations, and future directions regarding Archangel to highlight its
distinct value for the broader machine learning community.
Related papers
- Bayesian Detector Combination for Object Detection with Crowdsourced Annotations [49.43709660948812]
Acquiring fine-grained object detection annotations in unconstrained images is time-consuming, expensive, and prone to noise.
We propose a novel Bayesian Detector Combination (BDC) framework to more effectively train object detectors with noisy crowdsourced annotations.
BDC is model-agnostic, requires no prior knowledge of the annotators' skill level, and seamlessly integrates with existing object detection models.
arXiv Detail & Related papers (2024-07-10T18:00:54Z) - UEMM-Air: A Synthetic Multi-modal Dataset for Unmanned Aerial Vehicle Object Detection [14.869928980343415]
We propose a synthetic multi-modal UAV-based object detection dataset, UEMM-Air.
Specially, we simulate various UAV flight scenarios and object types using the Unreal Engine (UE)
In total, our UEMM-Air consists of 20k pairs of images with 5 modalities and precise annotations.
arXiv Detail & Related papers (2024-06-10T13:00:22Z) - MMAUD: A Comprehensive Multi-Modal Anti-UAV Dataset for Modern Miniature
Drone Threats [37.981623262267036]
MMAUD addresses a critical gap in contemporary threat detection methodologies by focusing on drone detection, UAV-type classification, and trajectory estimation.
It offers a unique overhead aerial detection vital for addressing real-world scenarios with higher fidelity than datasets captured on specific vantage points using thermal and RGB.
Our proposed modalities are cost-effective and highly adaptable, allowing users to experiment and implement new UAV threat detection tools.
arXiv Detail & Related papers (2024-02-06T04:57:07Z) - Diffusion-Based Particle-DETR for BEV Perception [94.88305708174796]
Bird-Eye-View (BEV) is one of the most widely-used scene representations for visual perception in Autonomous Vehicles (AVs)
Recent diffusion-based methods offer a promising approach to uncertainty modeling for visual perception but fail to effectively detect small objects in the large coverage of the BEV.
Here, we address this problem by combining the diffusion paradigm with current state-of-the-art 3D object detectors in BEV.
arXiv Detail & Related papers (2023-12-18T09:52:14Z) - Innovative Horizons in Aerial Imagery: LSKNet Meets DiffusionDet for
Advanced Object Detection [55.2480439325792]
We present an in-depth evaluation of an object detection model that integrates the LSKNet backbone with the DiffusionDet head.
The proposed model achieves a mean average precision (MAP) of approximately 45.7%, which is a significant improvement.
This advancement underscores the effectiveness of the proposed modifications and sets a new benchmark in aerial image analysis.
arXiv Detail & Related papers (2023-11-21T19:49:13Z) - UAV-Sim: NeRF-based Synthetic Data Generation for UAV-based Perception [62.71374902455154]
We leverage recent advancements in neural rendering to improve static and dynamic novelview UAV-based image rendering.
We demonstrate a considerable performance boost when a state-of-the-art detection model is optimized primarily on hybrid sets of real and synthetic data.
arXiv Detail & Related papers (2023-10-25T00:20:37Z) - Long Range Object-Level Monocular Depth Estimation for UAVs [0.0]
We propose several novel extensions to state-of-the-art methods for monocular object detection from images at long range.
Firstly, we propose Sigmoid and ReLU-like encodings when modeling depth estimation as a regression task.
Secondly, we frame the depth estimation as a classification problem and introduce a Soft-Argmax function in the calculation of the training loss.
arXiv Detail & Related papers (2023-02-17T15:26:04Z) - Vision-Based UAV Self-Positioning in Low-Altitude Urban Environments [20.69412701553767]
Unmanned Aerial Vehicles (UAVs) rely on satellite systems for stable positioning.
In such situations, vision-based techniques can serve as an alternative, ensuring the self-positioning capability of UAVs.
This paper presents a new dataset, DenseUAV, which is the first publicly available dataset designed for the UAV self-positioning task.
arXiv Detail & Related papers (2022-01-23T07:18:55Z) - Leveraging Synthetic Data in Object Detection on Unmanned Aerial
Vehicles [14.853897011640022]
We extend the open-source framework DeepGTAV to work for UAV scenarios.
We capture various large-scale high-resolution synthetic data sets in several domains to demonstrate their use in real-world object detection from UAVs.
arXiv Detail & Related papers (2021-12-22T22:41:02Z) - Cycle and Semantic Consistent Adversarial Domain Adaptation for Reducing
Simulation-to-Real Domain Shift in LiDAR Bird's Eye View [110.83289076967895]
We present a BEV domain adaptation method based on CycleGAN that uses prior semantic classification in order to preserve the information of small objects of interest during the domain adaptation process.
The quality of the generated BEVs has been evaluated using a state-of-the-art 3D object detection framework at KITTI 3D Object Detection Benchmark.
arXiv Detail & Related papers (2021-04-22T12:47:37Z) - Perceiving Traffic from Aerial Images [86.994032967469]
We propose an object detection method called Butterfly Detector that is tailored to detect objects in aerial images.
We evaluate our Butterfly Detector on two publicly available UAV datasets (UAVDT and VisDrone 2019) and show that it outperforms previous state-of-the-art methods while remaining real-time.
arXiv Detail & Related papers (2020-09-16T11:37:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.