Extending One-Stage Detection with Open-World Proposals
- URL: http://arxiv.org/abs/2201.02302v1
- Date: Fri, 7 Jan 2022 02:29:09 GMT
- Title: Extending One-Stage Detection with Open-World Proposals
- Authors: Sachin Konan and Kevin J Liang and Li Yin
- Abstract summary: We show that fully convolutional one-stage detection network FCOS can increase OWP performance by as much as 6% in recall on novel classes.
While two-stage methods worsen by 6% in recall on novel classes, we show that FCOS only drops 2% when jointly optimizing for OWP and classification.
- Score: 8.492340530784697
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many applications, such as autonomous driving, hand manipulation, or robot
navigation, object detection methods must be able to detect objects unseen in
the training set. Open World Detection(OWD) seeks to tackle this problem by
generalizing detection performance to seen and unseen class categories. Recent
works have seen success in the generation of class-agnostic proposals, which we
call Open-World Proposals(OWP), but this comes at the cost of a big drop on the
classification task when both tasks are considered in the detection model.
These works have investigated two-stage Region Proposal Networks (RPN) by
taking advantage of objectness scoring cues; however, for its simplicity,
run-time, and decoupling of localization and classification, we investigate OWP
through the lens of fully convolutional one-stage detection network, such as
FCOS. We show that our architectural and sampling optimizations on FCOS can
increase OWP performance by as much as 6% in recall on novel classes, marking
the first proposal-free one-stage detection network to achieve comparable
performance to RPN-based two-stage networks. Furthermore, we show that the
inherent, decoupled architecture of FCOS has benefits to retaining
classification performance. While two-stage methods worsen by 6% in recall on
novel classes, we show that FCOS only drops 2% when jointly optimizing for OWP
and classification.
Related papers
- LAA-Net: Localized Artifact Attention Network for Quality-Agnostic and Generalizable Deepfake Detection [12.567069964305265]
This paper introduces a novel approach for high-quality deepfake detection called Localized Artifact Attention Network (LAA-Net)
Existing methods for high-quality deepfake detection are mainly based on a supervised binary classifier coupled with an implicit attention mechanism.
Experiments performed on several benchmarks show the superiority of our approach in terms of Area Under the Curve (AUC) and Average Precision (AP)
arXiv Detail & Related papers (2024-01-24T23:42:08Z) - You Should Look at All Objects [28.862053913000384]
This paper revisits FPN in the detection framework and reveals the nature of the success of FPN from the perspective of optimization.
The degraded performance of large-scale objects is due to the arising of improper back-propagation paths after integrating FPN.
Two feasible strategies are proposed to enable each level of the backbone to look at all objects in the FPN-based detection frameworks.
arXiv Detail & Related papers (2022-07-16T09:59:34Z) - Distribution Alignment: A Unified Framework for Long-tail Visual
Recognition [52.36728157779307]
We propose a unified distribution alignment strategy for long-tail visual recognition.
We then introduce a generalized re-weight method in the two-stage learning to balance the class prior.
Our approach achieves the state-of-the-art results across all four recognition tasks with a simple and unified framework.
arXiv Detail & Related papers (2021-03-30T14:09:53Z) - Modulating Localization and Classification for Harmonized Object
Detection [40.82723262074911]
We propose a mutual learning framework to modulate the two tasks.
In particular, the two tasks are forced to learn from each other with a novel mutual labeling strategy.
We achieve a significant performance gain over the baseline detectors on the COCO dataset.
arXiv Detail & Related papers (2021-03-16T10:36:02Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z) - Corner Proposal Network for Anchor-free, Two-stage Object Detection [174.59360147041673]
The goal of object detection is to determine the class and location of objects in an image.
This paper proposes a novel anchor-free, two-stage framework which first extracts a number of object proposals.
We demonstrate that these two stages are effective solutions for improving recall and precision.
arXiv Detail & Related papers (2020-07-27T19:04:57Z) - FCOS: A simple and strong anchor-free object detector [111.87691210818194]
We propose a fully convolutional one-stage object detector (FCOS) to solve object detection in a per-pixel prediction fashion.
Almost all state-of-the-art object detectors such as RetinaNet, SSD, YOLOv3, and Faster R-CNN rely on pre-defined anchor boxes.
In contrast, our proposed detector FCOS is anchor box free, as well as proposal free.
arXiv Detail & Related papers (2020-06-14T01:03:39Z) - Fine-Grained Visual Classification with Efficient End-to-end
Localization [49.9887676289364]
We present an efficient localization module that can be fused with a classification network in an end-to-end setup.
We evaluate the new model on the three benchmark datasets CUB200-2011, Stanford Cars and FGVC-Aircraft.
arXiv Detail & Related papers (2020-05-11T14:07:06Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.