Wise-IoU: Bounding Box Regression Loss with Dynamic Focusing Mechanism
- URL: http://arxiv.org/abs/2301.10051v3
- Date: Sat, 8 Apr 2023 13:58:40 GMT
- Title: Wise-IoU: Bounding Box Regression Loss with Dynamic Focusing Mechanism
- Authors: Zanjia Tong, Yuhang Chen, Zewei Xu, Rong Yu
- Abstract summary: We propose an IoU-based loss with a dynamic non-monotonic FM named Wise-IoU (WIoU)
This strategy reduces the competitiveness of high-quality anchor boxes while also reducing the harmful gradient generated by low-quality examples.
When WIoU is applied to the state-of-the-art real-time detector YOLOv7, the AP-75 on the MS-COCO dataset is improved from 53.03% to 54.50%.
- Score: 7.645166402471877
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The loss function for bounding box regression (BBR) is essential to object
detection. Its good definition will bring significant performance improvement
to the model. Most existing works assume that the examples in the training data
are high-quality and focus on strengthening the fitting ability of BBR loss. If
we blindly strengthen BBR on low-quality examples, it will jeopardize
localization performance. Focal-EIoU v1 was proposed to solve this problem, but
due to its static focusing mechanism (FM), the potential of non-monotonic FM
was not fully exploited. Based on this idea, we propose an IoU-based loss with
a dynamic non-monotonic FM named Wise-IoU (WIoU). The dynamic non-monotonic FM
uses the outlier degree instead of IoU to evaluate the quality of anchor boxes
and provides a wise gradient gain allocation strategy. This strategy reduces
the competitiveness of high-quality anchor boxes while also reducing the
harmful gradient generated by low-quality examples. This allows WIoU to focus
on ordinary-quality anchor boxes and improve the detector's overall
performance. When WIoU is applied to the state-of-the-art real-time detector
YOLOv7, the AP-75 on the MS-COCO dataset is improved from 53.03% to 54.50%.
Code is available at https://github.com/Instinct323/wiou.
Related papers
- Unified-IoU: For High-Quality Object Detection [1.62877896907106]
We propose a new IoU loss function, called Unified-IoU (UIoU), which is more concerned with the weight assignment between different quality prediction boxes.
Our proposed method achieves better performance on multiple datasets, especially at a high IoU threshold.
arXiv Detail & Related papers (2024-08-13T04:56:45Z) - Attention-Driven Training-Free Efficiency Enhancement of Diffusion Models [29.863953001061635]
Diffusion Models (DMs) have exhibited superior performance in generating high-quality and diverse images.
Existing works mainly adopt a retraining process to enhance DM efficiency.
We introduce the Attention-driven Training-free Efficient Diffusion Model (AT-EDM) framework that leverages attention maps to perform run-time pruning of redundant tokens.
arXiv Detail & Related papers (2024-05-08T17:56:47Z) - Towards Robust Federated Learning via Logits Calibration on Non-IID Data [49.286558007937856]
Federated learning (FL) is a privacy-preserving distributed management framework based on collaborative model training of distributed devices in edge networks.
Recent studies have shown that FL is vulnerable to adversarial examples, leading to a significant drop in its performance.
In this work, we adopt the adversarial training (AT) framework to improve the robustness of FL models against adversarial example (AE) attacks.
arXiv Detail & Related papers (2024-03-05T09:18:29Z) - Inner-IoU: More Effective Intersection over Union Loss with Auxiliary
Bounding Box [10.03001043843768]
We propose Inner-IoU loss, which calculates IoU loss through auxiliary bounding boxes.
For different datasets and detectors, we introduce a scaling factor ratio to control the scale size of the auxiliary bounding boxes.
Finally, integrate Inner-IoU into the existing IoU-based loss functions for simulation and comparative experiments.
arXiv Detail & Related papers (2023-11-06T05:14:24Z) - FedNAR: Federated Optimization with Normalized Annealing Regularization [54.42032094044368]
We explore the choices of weight decay and identify that weight decay value appreciably influences the convergence of existing FL algorithms.
We develop Federated optimization with Normalized Annealing Regularization (FedNAR), a plug-in that can be seamlessly integrated into any existing FL algorithms.
arXiv Detail & Related papers (2023-10-04T21:11:40Z) - The KFIoU Loss for Rotated Object Detection [115.334070064346]
In this paper, we argue that one effective alternative is to devise an approximate loss who can achieve trend-level alignment with SkewIoU loss.
Specifically, we model the objects as Gaussian distribution and adopt Kalman filter to inherently mimic the mechanism of SkewIoU.
The resulting new loss called KFIoU is easier to implement and works better compared with exact SkewIoU.
arXiv Detail & Related papers (2022-01-29T10:54:57Z) - A Systematic IoU-Related Method: Beyond Simplified Regression for Better
Localization [9.036025934093965]
We propose a new metric, the extended IoU, which is well-defined when two boxes are not overlapping and reduced to the standard IoU when overlapping.
Thirdly, we propose a steady optimization technique (SOT) to make the fractional EIoU loss approaching the minimum more steadily and smoothly.
arXiv Detail & Related papers (2021-12-03T09:00:55Z) - Alpha-IoU: A Family of Power Intersection over Union Losses for Bounding
Box Regression [59.72580239998315]
We generalize existing IoU-based losses to a new family of power IoU losses that have a power IoU term and an additional power regularization term.
Experiments on multiple object detection benchmarks and models demonstrate that $alpha$-IoU losses can surpass existing IoU-based losses by a noticeable performance margin.
arXiv Detail & Related papers (2021-10-26T13:09:20Z) - Focal and Efficient IOU Loss for Accurate Bounding Box Regression [63.14659624634066]
In object detection, bounding box regression (BBR) is a crucial step that determines the object localization performance.
Most previous loss functions for BBR have two main drawbacks: (i) Both $ell_n$-norm and IOU-based loss functions are inefficient to depict the objective of BBR, which leads to slow convergence and inaccurate regression results.
arXiv Detail & Related papers (2021-01-20T14:33:58Z) - Dynamic R-CNN: Towards High Quality Object Detection via Dynamic
Training [70.2914594796002]
We propose Dynamic R-CNN to adjust the label assignment criteria and the shape of regression loss function.
Our method improves upon ResNet-50-FPN baseline with 1.9% AP and 5.5% AP$_90$ on the MS dataset with no extra overhead.
arXiv Detail & Related papers (2020-04-13T15:20:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.