Combining Visible and Infrared Spectrum Imagery using Machine Learning
for Small Unmanned Aerial System Detection
- URL: http://arxiv.org/abs/2003.12638v2
- Date: Thu, 2 Apr 2020 19:08:02 GMT
- Title: Combining Visible and Infrared Spectrum Imagery using Machine Learning
for Small Unmanned Aerial System Detection
- Authors: Vinicius G. Goecks, Grayson Woods, John Valasek
- Abstract summary: This research work proposes combining the advantages of the LWIR and visible spectrum sensors using machine learning for vision-based detection of sUAS.
Our approach achieved a detection rate of 71.2 +- 8.3%, improving by 69% when compared to LWIR and by 30.4% when visible spectrum alone.
- Score: 1.392250707100996
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Advances in machine learning and deep neural networks for object detection,
coupled with lower cost and power requirements of cameras, led to promising
vision-based solutions for sUAS detection. However, solely relying on the
visible spectrum has previously led to reliability issues in low contrast
scenarios such as sUAS flying below the treeline and against bright sources of
light. Alternatively, due to the relatively high heat signatures emitted from
sUAS during flight, a long-wave infrared (LWIR) sensor is able to produce
images that clearly contrast the sUAS from its background. However, compared to
widely available visible spectrum sensors, LWIR sensors have lower resolution
and may produce more false positives when exposed to birds or other heat
sources. This research work proposes combining the advantages of the LWIR and
visible spectrum sensors using machine learning for vision-based detection of
sUAS. Utilizing the heightened background contrast from the LWIR sensor
combined and synchronized with the relatively increased resolution of the
visible spectrum sensor, a deep learning model was trained to detect the sUAS
through previously difficult environments. More specifically, the approach
demonstrated effective detection of multiple sUAS flying above and below the
treeline, in the presence of heat sources, and glare from the sun. Our approach
achieved a detection rate of 71.2 +- 8.3%, improving by 69% when compared to
LWIR and by 30.4% when visible spectrum alone, and achieved false alarm rate of
2.7 +- 2.6%, decreasing by 74.1% and by 47.1% when compared to LWIR and visible
spectrum alone, respectively, on average, for single and multiple drone
scenarios, controlled for the same confidence metric of the machine learning
object detector of at least 50%. Videos of the solution's performance can be
seen at https://sites.google.com/view/tamudrone-spie2020/.
Related papers
- Beyond Night Visibility: Adaptive Multi-Scale Fusion of Infrared and
Visible Images [49.75771095302775]
We propose an Adaptive Multi-scale Fusion network (AMFusion) with infrared and visible images.
First, we separately fuse spatial and semantic features from infrared and visible images, where the former are used for the adjustment of light distribution.
Second, we utilize detection features extracted by a pre-trained backbone that guide the fusion of semantic features.
Third, we propose a new illumination loss to constrain fusion image with normal light intensity.
arXiv Detail & Related papers (2024-03-02T03:52:07Z) - DensePose From WiFi [86.61881052177228]
We develop a deep neural network that maps the phase and amplitude of WiFi signals to UV coordinates within 24 human regions.
Our model can estimate the dense pose of multiple subjects, with comparable performance to image-based approaches.
arXiv Detail & Related papers (2022-12-31T16:48:43Z) - Near-infrared and visible-light periocular recognition with Gabor
features using frequency-adaptive automatic eye detection [69.35569554213679]
Periocular recognition has gained attention recently due to demands of increased robustness of face or iris in less controlled scenarios.
We present a new system for eye detection based on complex symmetry filters, which has the advantage of not needing training.
This system is used as input to a periocular algorithm based on retinotopic sampling grids and Gabor spectrum decomposition.
arXiv Detail & Related papers (2022-11-10T13:04:03Z) - Target-aware Dual Adversarial Learning and a Multi-scenario
Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection [65.30079184700755]
This study addresses the issue of fusing infrared and visible images that appear differently for object detection.
Previous approaches discover commons underlying the two modalities and fuse upon the common space either by iterative optimization or deep networks.
This paper proposes a bilevel optimization formulation for the joint problem of fusion and detection, and then unrolls to a target-aware Dual Adversarial Learning (TarDAL) network for fusion and a commonly used detection network.
arXiv Detail & Related papers (2022-03-30T11:44:56Z) - Long-Range Thermal 3D Perception in Low Contrast Environments [0.0]
This report discusses the results of Phase I effort to prove the feasibility of dramatic improvement of the microbolometer-based Long Wave Infrared (LWIR) detectors sensitivity.
The resulting low SWaP-C thermal depth-sensing system will enable the situational awareness of Autonomous Air Vehicles for Advanced Air Mobility (AAM)
It will provide robust 3D information of the surrounding environment, including low-contrast static and moving objects, at far distances in degraded visual conditions and GPS-denied areas.
arXiv Detail & Related papers (2021-12-10T01:16:44Z) - A Multi-Stage model based on YOLOv3 for defect detection in PV panels
based on IR and Visible Imaging by Unmanned Aerial Vehicle [65.99880594435643]
We propose a novel model to detect panel defects on aerial images captured by unmanned aerial vehicle.
The model combines detections of panels and defects to refine its accuracy.
The proposed model has been validated on two big PV plants in the south of Italy.
arXiv Detail & Related papers (2021-11-23T08:04:32Z) - Deep Learning-based Human Detection for UAVs with Optical and Infrared
Cameras: System and Experiments [35.342730238802886]
We present our deep learning-based human detection system that uses optical (RGB) and long-wave infrared (LWIR) cameras.
In each spectrum, a customized RetinaNet network with ResNet backbone provides human detections which are subsequently fused to minimize the overall false detection rate.
We show that by optimizing the bounding box anchors and augmenting the image resolution the number of missed detections from high altitudes can be decreased by over 20 percent.
arXiv Detail & Related papers (2020-08-10T15:30:42Z) - Recurrent Exposure Generation for Low-Light Face Detection [113.25331155337759]
We propose a novel Recurrent Exposure Generation (REG) module and a Multi-Exposure Detection (MED) module.
REG produces progressively and efficiently intermediate images corresponding to various exposure settings.
Such pseudo-exposures are then fused by MED to detect faces across different lighting conditions.
arXiv Detail & Related papers (2020-07-21T17:30:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.