Lightweight Multi-Drone Detection and 3D-Localization via YOLO
- URL: http://arxiv.org/abs/2202.09097v1
- Date: Fri, 18 Feb 2022 09:41:23 GMT
- Title: Lightweight Multi-Drone Detection and 3D-Localization via YOLO
- Authors: Aryan Sharma, Nitik Jain, and Mangal Kothari
- Abstract summary: We present and evaluate a method to perform real-time multiple drone detection and three-dimensional localization.
We use state-of-the-art tiny-YOLOv4 object detection algorithm and stereo triangulation.
Our computer vision approach eliminates the need for computationally expensive stereo matching algorithms.
- Score: 1.284647943889634
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we present and evaluate a method to perform real-time multiple
drone detection and three-dimensional localization using state-of-the-art
tiny-YOLOv4 object detection algorithm and stereo triangulation. Our computer
vision approach eliminates the need for computationally expensive stereo
matching algorithms, thereby significantly reducing the memory footprint and
making it deployable on embedded systems. Our drone detection system is highly
modular (with support for various detection algorithms) and capable of
identifying multiple drones in a system, with real-time detection accuracy of
up to 77\% with an average FPS of 332 (on Nvidia Titan Xp). We also test the
complete pipeline in AirSim environment, detecting drones at a maximum distance
of 8 meters, with a mean error of $23\%$ of the distance. We also release the
source code for the project, with pre-trained models and the curated synthetic
stereo dataset.
Related papers
- C2FDrone: Coarse-to-Fine Drone-to-Drone Detection using Vision Transformer Networks [23.133250476580038]
A vision-based drone-to-drone detection system is crucial for various applications like collision avoidance, countering hostile drones, and search-and-rescue operations.
detecting drones presents unique challenges, including small object sizes, distortion, and real-time processing requirements.
We propose a novel coarse-to-fine detection strategy based on vision transformers.
arXiv Detail & Related papers (2024-04-30T05:51:21Z) - Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - An Empirical Analysis of Range for 3D Object Detection [70.54345282696138]
We present an empirical analysis of far-field 3D detection using the long-range detection dataset Argoverse 2.0.
Near-field LiDAR measurements are dense and optimally encoded by small voxels, while far-field measurements are sparse and are better encoded with large voxels.
We propose simple techniques to efficiently ensemble models for long-range detection that improve efficiency by 33% and boost accuracy by 3.2% CDS.
arXiv Detail & Related papers (2023-08-08T05:29:26Z) - TransVisDrone: Spatio-Temporal Transformer for Vision-based
Drone-to-Drone Detection in Aerial Videos [57.92385818430939]
Drone-to-drone detection using visual feed has crucial applications, such as detecting drone collisions, detecting drone attacks, or coordinating flight with other drones.
Existing methods are computationally costly, follow non-end-to-end optimization, and have complex multi-stage pipelines, making them less suitable for real-time deployment on edge devices.
We propose a simple yet effective framework, itTransVisDrone, that provides an end-to-end solution with higher computational efficiency.
arXiv Detail & Related papers (2022-10-16T03:05:13Z) - Small Object Detection using Deep Learning [0.28675177318965034]
The proposed system consists of a custom deep learning model Tiny YOLOv3, one of the flavors of very fast object detection model You Look Only Once (YOLO) is built and used for detection.
The proposed architecture has shown significantly better performance as compared to the previous YOLO version.
arXiv Detail & Related papers (2022-01-10T09:58:25Z) - CFTrack: Center-based Radar and Camera Fusion for 3D Multi-Object
Tracking [9.62721286522053]
We propose an end-to-end network for joint object detection and tracking based on radar and camera sensor fusion.
Our proposed method uses a center-based radar-camera fusion algorithm for object detection and utilizes a greedy algorithm for object association.
We evaluate our method on the challenging nuScenes dataset, where it achieves 20.0 AMOTA and outperforms all vision-based 3D tracking methods in the benchmark.
arXiv Detail & Related papers (2021-07-11T23:56:53Z) - Detection, Tracking, and Counting Meets Drones in Crowds: A Benchmark [97.07865343576361]
We construct a benchmark with a new drone-captured largescale dataset, named as DroneCrowd.
We annotate 20,800 people trajectories with 4.8 million heads and several video-level attributes.
We design the Space-Time Neighbor-Aware Network (STNNet) as a strong baseline to solve object detection, tracking and counting jointly in dense crowds.
arXiv Detail & Related papers (2021-05-06T04:46:14Z) - Dogfight: Detecting Drones from Drones Videos [58.158988162743825]
This paper attempts to address the problem of drones detection from other flying drones variations.
The erratic movement of the source and target drones, small size, arbitrary shape, large intensity, and occlusion make this problem quite challenging.
To handle this, instead of using region-proposal based methods, we propose to use a two-stage segmentation-based approach.
arXiv Detail & Related papers (2021-03-31T17:43:31Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Expedited Multi-Target Search with Guaranteed Performance via
Multi-fidelity Gaussian Processes [9.434133337939496]
We consider a scenario in which an autonomous vehicle operates in a 3D environment and is tasked with searching for an unknown number of stationary targets on the 2D floor of the environment.
We model the sensing field using a multi-fidelity Gaussian process that systematically describes the sensing information available at different altitudes from the floor.
Based on the sensing model, we design a novel algorithm called Multi-Target Search (EMTS) that addresses the coverage-accuracy trade-off.
arXiv Detail & Related papers (2020-05-18T02:53:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.