Being Aware of Localization Accuracy By Generating Predicted-IoU-Guided
Quality Scores
- URL: http://arxiv.org/abs/2309.13269v1
- Date: Sat, 23 Sep 2023 05:27:59 GMT
- Title: Being Aware of Localization Accuracy By Generating Predicted-IoU-Guided
Quality Scores
- Authors: Pengfei Liu, Weibo Wang, Yuhan Guo, Jiubin Tan
- Abstract summary: We develop an elegant LQE branch to acquire localization quality score guided by predicted IoU.
A novel one stage detector termed CLQ is proposed.
Experiments show that CLQ achieves state-of-the-arts' performance at an accuracy of 47.8 AP and a speed of 11.5 fps.
- Score: 24.086202809990795
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Localization Quality Estimation (LQE) helps to improve detection performance
as it benefits post processing through jointly considering classification score
and localization accuracy. In this perspective, for further leveraging the
close relationship between localization accuracy and IoU
(Intersection-Over-Union), and for depressing those inconsistent predictions,
we designed an elegant LQE branch to acquire localization quality score guided
by predicted IoU. Distinctly, for alleviating the inconsistency of
classification score and localization quality during training and inference,
under which some predictions with low classification scores but high LQE scores
will impair the performance, instead of separately and independently setting,
we embedded LQE branch into classification branch, producing a joint
classification-localization-quality representation. Then a novel one stage
detector termed CLQ is proposed. Extensive experiments show that CLQ achieves
state-of-the-arts' performance at an accuracy of 47.8 AP and a speed of 11.5
fps with ResNeXt-101 as backbone on COCO test-dev. Finally, we extend CLQ to
ATSS, producing a reliable 1.2 AP gain, showing our model's strong adaptability
and scalability. Codes are released at https://github.com/PanffeeReal/CLQ.
Related papers
- Token-based Decision Criteria Are Suboptimal in In-context Learning [2.2973949268669562]
In-Context Learning (ICL) typically utilizes classification criteria from output probabilities of manually selected label tokens.
We argue that such token-based classification criteria lead to suboptimal decision boundaries.
We propose Hidden, which renounces token probabilities and uses the nearest centroid on the LM's last hidden states.
arXiv Detail & Related papers (2024-06-24T11:16:26Z) - Rank-DETR for High Quality Object Detection [52.82810762221516]
A highly performant object detector requires accurate ranking for the bounding box predictions.
In this work, we introduce a simple and highly performant DETR-based object detector by proposing a series of rank-oriented designs.
arXiv Detail & Related papers (2023-10-13T04:48:32Z) - ProTeCt: Prompt Tuning for Taxonomic Open Set Classification [59.59442518849203]
Few-shot adaptation methods do not fare well in the taxonomic open set (TOS) setting.
We propose a prompt tuning technique that calibrates the hierarchical consistency of model predictions.
A new Prompt Tuning for Hierarchical Consistency (ProTeCt) technique is then proposed to calibrate classification across label set granularities.
arXiv Detail & Related papers (2023-06-04T02:55:25Z) - AUC-based Selective Classification [5.406386303264086]
We propose a model-agnostic approach to associate a selection function to a given binary classifier.
We provide both theoretical justifications and a novel algorithm, called $AUCross$, to achieve such a goal.
Experiments show that $AUCross$ succeeds in trading-off coverage for AUC, improving over existing selective classification methods targeted at optimizing accuracy.
arXiv Detail & Related papers (2022-10-19T16:29:50Z) - Learning Localization-aware Target Confidence for Siamese Visual
Tracking [13.684278662495204]
We propose a novel tracking paradigm, called SiamLA.
Within this paradigm, a series of simple, yet effective localization-aware components are introduced.
Our SiamLA achieves state-of-the-art performance in terms of both accuracy and efficiency.
arXiv Detail & Related papers (2022-04-29T13:37:15Z) - Generalized Focal Loss V2: Learning Reliable Localization Quality
Estimation for Dense Object Detection [78.11775981796367]
GFLV2 (ResNet-101) achieves 46.2 AP at 14.6 FPS, surpassing the previous state-of-the-art ATSS baseline (43.6 AP at 14.6 FPS) by absolute 2.6 AP on COCO tt test-dev.
Code will be available at https://github.com/implus/GFocalV2.
arXiv Detail & Related papers (2020-11-25T17:06:37Z) - Generalized Focal Loss: Learning Qualified and Distributed Bounding
Boxes for Dense Object Detection [85.53263670166304]
One-stage detector basically formulates object detection as dense classification and localization.
Recent trend for one-stage detectors is to introduce an individual prediction branch to estimate the quality of localization.
This paper delves into the representations of the above three fundamental elements: quality estimation, classification and localization.
arXiv Detail & Related papers (2020-06-08T07:24:33Z) - Fine-Grained Visual Classification with Efficient End-to-end
Localization [49.9887676289364]
We present an efficient localization module that can be fused with a classification network in an end-to-end setup.
We evaluate the new model on the three benchmark datasets CUB200-2011, Stanford Cars and FGVC-Aircraft.
arXiv Detail & Related papers (2020-05-11T14:07:06Z) - Scope Head for Accurate Localization in Object Detection [135.9979405835606]
We propose a novel detector coined as ScopeNet, which models anchors of each location as a mutually dependent relationship.
With our concise and effective design, the proposed ScopeNet achieves state-of-the-art results on COCO.
arXiv Detail & Related papers (2020-05-11T04:00:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.