N-EIoU-YOLOv9: A Signal-Aware Bounding Box Regression Loss for Lightweight Mobile Detection of Rice Leaf Diseases
- URL: http://arxiv.org/abs/2601.09170v1
- Date: Wed, 14 Jan 2026 05:13:36 GMT
- Title: N-EIoU-YOLOv9: A Signal-Aware Bounding Box Regression Loss for Lightweight Mobile Detection of Rice Leaf Diseases
- Authors: Dung Ta Nguyen Duc, Thanh Bui Dang, Hoang Le Minh, Tung Nguyen Viet, Huong Nguyen Thanh, Dong Trinh Cong,
- Abstract summary: We propose N EIoU YOLOv9, a lightweight detection framework based on a signal aware bounding box regression loss.<n>The proposed loss reshapes localization gradient by combining non monotonic focusing with decoupled width and height optimization.<n>This design is particularly effective for small and low contrast targets commonly observed in agricultural disease imagery.
- Score: 0.6280530476948474
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we propose N EIoU YOLOv9, a lightweight detection framework based on a signal aware bounding box regression loss derived from non monotonic gradient focusing and geometric decoupling principles, referred to as N EIoU (Non monotonic Efficient Intersection over Union). The proposed loss reshapes localization gradients by combining non monotonic focusing with decoupled width and height optimization, thereby enhancing weak regression signals for hard samples with low overlap while reducing gradient interference. This design is particularly effective for small and low contrast targets commonly observed in agricultural disease imagery. The proposed N EIoU loss is integrated into a lightweight YOLOv9t architecture and evaluated on a self collected field dataset comprising 5908 rice leaf images across four disease categories and healthy leaves. Experimental results demonstrate consistent performance gains over the standard CIoU loss, achieving a mean Average Precision of 90.3 percent, corresponding to a 4.3 percent improvement over the baseline, with improved localization accuracy under stricter evaluation criteria. For practical validation, the optimized model is deployed on an Android device using TensorFlow Lite with Float16 quantization, achieving an average inference time of 156 milliseconds per frame while maintaining accuracy. These results confirm that the proposed approach effectively balances accuracy, optimization stability, and computational efficiency for edge based agricultural monitoring systems.
Related papers
- BadCLIP++: Stealthy and Persistent Backdoors in Multimodal Contrastive Learning [73.46118996284888]
Research on backdoor attacks against multimodal contrastive learning models faces two key challenges: stealthiness and persistence.<n>We propose BadCLIP++, a unified framework that tackles both challenges.<n>For stealthiness, we introduce a semantic-fusion QR micro-trigger that embeds imperceptible patterns near task-relevant regions.<n>For persistence, we stabilize trigger embeddings via radius shrinkage and centroid alignment.
arXiv Detail & Related papers (2026-02-19T08:31:16Z) - Enhancing YOLOv11n for Reliable Child Detection in Noisy Surveillance Footage [0.0]
This paper presents a practical and lightweight solution for enhancing child detection in low-quality surveillance footage.<n>Building upon the efficient YOLOv11n architecture, we propose a deployment-ready pipeline that improves detection under challenging conditions.<n>The entire pipeline maintains compatibility with low-power edge devices and supports real-time performance.
arXiv Detail & Related papers (2026-02-11T07:27:47Z) - Smooth Operator: Smooth Verifiable Reward Activates Spatial Reasoning Ability of Vision-Language Model [18.526821056010384]
Vision-Language Models (VLMs) face a critical bottleneck in achieving precise numerical prediction for 3D scene understanding.<n>Traditional reinforcement learning approaches, primarily based on relative ranking, often suffer from severe reward sparsity and gradient instability.<n>We introduce the Smooth Numerical Reward Activation (SNRA) operator and the Absolute-Preserving GRPO framework.
arXiv Detail & Related papers (2026-01-12T16:26:42Z) - Noise-Robust Tiny Object Localization with Flows [63.60972031108944]
We propose a noise-robust localization framework leveraging normalizing flows for flexible error modeling and uncertainty-guided optimization.<n>Our method captures complex, non-Gaussian prediction distributions through flow-based error modeling, enabling robust learning under noisy supervision.<n>An uncertainty-aware gradient modulation mechanism further suppresses learning from high-uncertainty, noise-prone samples, mitigating overfitting while stabilizing training.
arXiv Detail & Related papers (2026-01-02T09:16:55Z) - TY-RIST: Tactical YOLO Tricks for Real-time Infrared Small Target Detection [6.0340092200636475]
Infrared small target detection (IRSTD) is critical for defense and surveillance but remains challenging.<n>We propose TY-RIST, an optimized YOLOv12n architecture that integrates a stride-aware backbone with fine-grained receptive fields.<n>Experiments on four benchmarks and across 20 different models demonstrate state-of-the-art performance, improving mAP at 0.5 IoU by +7.9%, Precision by +3%, and Recall by +10.2%.
arXiv Detail & Related papers (2025-09-26T20:36:57Z) - A Lightweight and Robust Framework for Real-Time Colorectal Polyp Detection Using LOF-Based Preprocessing and YOLO-v11n [2.4896276458398785]
This study introduces a new, lightweight, and efficient framework for polyp detection.<n>It combines the Local Outlier Factor algorithm for filtering noisy data with the YOLO-v11n deep learning model.<n>Compared to previous YOLO-based methods, our model demonstrates enhanced accuracy and efficiency.
arXiv Detail & Related papers (2025-07-14T23:36:54Z) - LAM-YOLO: Drones-based Small Object Detection on Lighting-Occlusion Attention Mechanism YOLO [0.9062164411594178]
LAM-YOLO is an object detection model specifically designed for drone-based images.
We introduce a light-occlusion attention mechanism to enhance the visibility of small targets under different lighting conditions.
Second, we utilize an improved SIB-IoU as the regression loss function to accelerate model convergence and enhance localization accuracy.
arXiv Detail & Related papers (2024-11-01T10:00:48Z) - SIRST-5K: Exploring Massive Negatives Synthesis with Self-supervised
Learning for Robust Infrared Small Target Detection [53.19618419772467]
Single-frame infrared small target (SIRST) detection aims to recognize small targets from clutter backgrounds.
With the development of Transformer, the scale of SIRST models is constantly increasing.
With a rich diversity of infrared small target data, our algorithm significantly improves the model performance and convergence speed.
arXiv Detail & Related papers (2024-03-08T16:14:54Z) - Align-DETR: Enhancing End-to-end Object Detection with Aligned Loss [35.11300328598727]
This paper identifies two key forms of misalignment within the model.<n>We introduce a novel loss function, termed as Align Loss, to resolve the discrepancy between the two tasks.<n>Our method achieves a 49.3% (+0.6) AP on the H-DETR baseline with the ResNet-50 backbone.
arXiv Detail & Related papers (2023-04-15T10:24:51Z) - The KFIoU Loss for Rotated Object Detection [115.334070064346]
In this paper, we argue that one effective alternative is to devise an approximate loss who can achieve trend-level alignment with SkewIoU loss.
Specifically, we model the objects as Gaussian distribution and adopt Kalman filter to inherently mimic the mechanism of SkewIoU.
The resulting new loss called KFIoU is easier to implement and works better compared with exact SkewIoU.
arXiv Detail & Related papers (2022-01-29T10:54:57Z) - SADet: Learning An Efficient and Accurate Pedestrian Detector [68.66857832440897]
This paper proposes a series of systematic optimization strategies for the detection pipeline of one-stage detector.
It forms a single shot anchor-based detector (SADet) for efficient and accurate pedestrian detection.
Though structurally simple, it presents state-of-the-art result and real-time speed of $20$ FPS for VGA-resolution images.
arXiv Detail & Related papers (2020-07-26T12:32:38Z) - Neural Control Variates [71.42768823631918]
We show that a set of neural networks can face the challenge of finding a good approximation of the integrand.
We derive a theoretically optimal, variance-minimizing loss function, and propose an alternative, composite loss for stable online training in practice.
Specifically, we show that the learned light-field approximation is of sufficient quality for high-order bounces, allowing us to omit the error correction and thereby dramatically reduce the noise at the cost of negligible visible bias.
arXiv Detail & Related papers (2020-06-02T11:17:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.