Multi-temporal Adaptive Red-Green-Blue and Long-Wave Infrared Fusion for You Only Look Once-Based Landmine Detection from Unmanned Aerial Systems
- URL: http://arxiv.org/abs/2512.20487v1
- Date: Tue, 23 Dec 2025 16:26:47 GMT
- Title: Multi-temporal Adaptive Red-Green-Blue and Long-Wave Infrared Fusion for You Only Look Once-Based Landmine Detection from Unmanned Aerial Systems
- Authors: James E. Gallagher, Edward J. Oughton, Jana Kosecka,
- Abstract summary: 110 million actively deployed mines across 60 countries, claiming 26,000 casualties annually.<n>This research evaluates adaptive Red-Green-Blue (RGB) and Long-Wave Infrared (LWIR) fusion for Unmanned Aerial Systems (UAS)-based detection of surface-laid landmines.
- Score: 2.2976554778751668
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Landmines remain a persistent humanitarian threat, with 110 million actively deployed mines across 60 countries, claiming 26,000 casualties annually. This research evaluates adaptive Red-Green-Blue (RGB) and Long-Wave Infrared (LWIR) fusion for Unmanned Aerial Systems (UAS)-based detection of surface-laid landmines, leveraging the thermal contrast between the ordnance and the surrounding soil to enhance feature extraction. Using You Only Look Once (YOLO) architectures (v8, v10, v11) across 114 test images, generating 35,640 model-condition evaluations, YOLOv11 achieved optimal performance (86.8% mAP), with 10 to 30% thermal fusion at 5 to 10m altitude identified as the optimal detection parameters. A complementary architectural comparison revealed that while RF-DETR achieved the highest accuracy (69.2% mAP), followed by Faster R-CNN (67.6%), YOLOv11 (64.2%), and RetinaNet (50.2%), YOLOv11 trained 17.7 times faster than the transformer-based RF-DETR (41 minutes versus 12 hours), presenting a critical accuracy-efficiency tradeoff for operational deployment. Aggregated multi-temporal training datasets outperformed season-specific approaches by 1.8 to 9.6%, suggesting that models benefit from exposure to diverse thermal conditions. Anti-Tank (AT) mines achieved 61.9% detection accuracy, compared with 19.2% for Anti-Personnel (AP) mines, reflecting both the size differential and thermal-mass differences between these ordnance classes. As this research examined surface-laid mines where thermal contrast is maximized, future research should quantify thermal contrast effects for mines buried at varying depths across heterogeneous soil types.
Related papers
- Detection of On-Ground Chestnuts Using Artificial Intelligence Toward Automated Picking [0.09176056742068812]
Traditional mechanized chestnut harvesting is too costly for small producers.<n> Accurate, reliable detection of chestnuts on the orchard floor is crucial for developing low-cost, vision-guided automated harvesting technology.<n>This study collected 319 images of chestnuts on the orchard floor, containing 6524 annotated chestnuts.
arXiv Detail & Related papers (2026-02-15T13:28:23Z) - AMLID: An Adaptive Multispectral Landmine Identification Dataset for Drone-Based Detection [0.0]
We present the first open-source dataset combining Red-Green-Blue (RGB) and Long-Wave Infrared (LWIR) imagery for Unmanned Aerial Systems (UAS)-based landmine detection.<n>AMLID comprises of 12,078 labeled images featuring 21 globally deployed landmine types across anti-personnel and anti-tank categories in both metal and plastic compositions.
arXiv Detail & Related papers (2025-12-21T13:58:35Z) - YOLOv11-Litchi: Efficient Litchi Fruit Detection based on UAV-Captured Agricultural Imagery in Complex Orchard Environments [6.862722449907841]
This paper introduces YOLOv11-Litchi, a lightweight and robust detection model specifically designed for UAV-based litchi detection.<n>YOLOv11-Litchi achieves a parameter size of 6.35 MB - 32.5% smaller than the YOLOv11 baseline.<n>The model achieves a frame rate of 57.2 FPS, meeting real-time detection requirements.
arXiv Detail & Related papers (2025-10-11T09:44:00Z) - A Comparative Benchmark of Real-time Detectors for Blueberry Detection towards Precision Orchard Management [2.667064587590596]
This study presents a novel comparative benchmark analysis of advanced real-time object detectors.<n>This dataset comprises 661 canopy images collected with smartphones during the 2022-2023 seasons.<n>Among the YOLO models, YOLOv12m achieved the best accuracy with a mAP@50 of 93.3%.
arXiv Detail & Related papers (2025-09-24T21:42:24Z) - Solar Active Regions Emergence Prediction Using Long Short-Term Memory
Networks [44.99833362998488]
We develop Long Short-Term Memory (LSTM) models to predict the formation of active regions (ARs) on the solar surface.
Time-series datasets of acoustic power and magnetic flux are used to train LSTM models on predicting continuum intensity, 12 hours in advance.
These novel machine learning (ML) models are able to capture variations of the acoustic power density associated with upcoming magnetic flux emergence and continuum intensity decrease.
arXiv Detail & Related papers (2024-09-25T23:09:46Z) - Thermal Imaging and Radar for Remote Sleep Monitoring of Breathing and Apnea [42.00356210257671]
We show the first comparison of radar and thermal imaging for sleep monitoring.
Our thermal imaging method detects apneas with an accuracy of 0.99, a precision of 0.68, a recall of 0.74, an F1 score of 0.71, and an intra-class correlation of 0.73.
We present a novel proposal for classifying obstructive and central sleep apnea by leveraging a multimodal setup.
arXiv Detail & Related papers (2024-07-16T17:26:50Z) - Dual-band feature selection for maturity classification of specialty crops by hyperspectral imaging [5.905721043072562]
The maturity classification of specialty crops such as strawberries and tomatoes is an essential agricultural downstream activity.
Recent advancements in Deep Learning have produced encouraging results in color images for maturity classification applications.
We present a feature extraction method to compute yet distinctive features for the maturity classification.
arXiv Detail & Related papers (2024-05-16T10:01:16Z) - Two Scalable Approaches for Burned-Area Mapping Using U-Net and Landsat
Imagery [39.91303506884272]
This study explores two proposed approaches based on the U-Net model for automating and optimizing the burned-area mapping process.
Tests based on 195 representative images of the study area show that increasing dataset balance using the AS model yields better performance.
arXiv Detail & Related papers (2023-11-29T05:42:25Z) - Automating global landslide detection with heterogeneous ensemble
deep-learning classification [44.99833362998488]
Landslides threaten infrastructure, including roads, railways, buildings, and human life.
Hazard-based spatial planning and early warning systems are cost-effective strategies to reduce the risk to society from landslides.
Deep learning models have recently been applied for landside mapping using medium- to high-resolution satellite images as input.
arXiv Detail & Related papers (2023-09-12T10:56:16Z) - Vision Transformers, a new approach for high-resolution and large-scale
mapping of canopy heights [50.52704854147297]
We present a new vision transformer (ViT) model optimized with a classification (discrete) and a continuous loss function.
This model achieves better accuracy than previously used convolutional based approaches (ConvNets) optimized with only a continuous loss function.
arXiv Detail & Related papers (2023-04-22T22:39:03Z) - A Multi-Stage model based on YOLOv3 for defect detection in PV panels
based on IR and Visible Imaging by Unmanned Aerial Vehicle [65.99880594435643]
We propose a novel model to detect panel defects on aerial images captured by unmanned aerial vehicle.
The model combines detections of panels and defects to refine its accuracy.
The proposed model has been validated on two big PV plants in the south of Italy.
arXiv Detail & Related papers (2021-11-23T08:04:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.