Towards Discriminative and Transferable One-Stage Few-Shot Object
Detectors
- URL: http://arxiv.org/abs/2210.05783v1
- Date: Tue, 11 Oct 2022 20:58:25 GMT
- Title: Towards Discriminative and Transferable One-Stage Few-Shot Object
Detectors
- Authors: Karim Guirguis, Mohamed Abdelsamad, George Eskandar, Ahmed Hendawy,
Matthias Kayser, Bin Yang, Juergen Beyerer
- Abstract summary: Few-shot object detection (FSOD) aims to address this problem by learning novel classes given only a few samples.
We make the observation that the large gap in performance between two-stage and one-stage FSODs are mainly due to their weak discriminability.
To address these limitations, we propose the Few-shot RetinaNet (FSRN) that consists of: a multi-way support training strategy to augment the number of foreground samples for dense meta-detectors.
- Score: 3.9189402702217344
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent object detection models require large amounts of annotated data for
training a new classes of objects. Few-shot object detection (FSOD) aims to
address this problem by learning novel classes given only a few samples. While
competitive results have been achieved using two-stage FSOD detectors,
typically one-stage FSODs underperform compared to them. We make the
observation that the large gap in performance between two-stage and one-stage
FSODs are mainly due to their weak discriminability, which is explained by a
small post-fusion receptive field and a small number of foreground samples in
the loss function. To address these limitations, we propose the Few-shot
RetinaNet (FSRN) that consists of: a multi-way support training strategy to
augment the number of foreground samples for dense meta-detectors, an early
multi-level feature fusion providing a wide receptive field that covers the
whole anchor area and two augmentation techniques on query and source images to
enhance transferability. Extensive experiments show that the proposed approach
addresses the limitations and boosts both discriminability and transferability.
FSRN is almost two times faster than two-stage FSODs while remaining
competitive in accuracy, and it outperforms the state-of-the-art of one-stage
meta-detectors and also some two-stage FSODs on the MS-COCO and PASCAL VOC
benchmarks.
Related papers
- Efficient Meta-Learning Enabled Lightweight Multiscale Few-Shot Object Detection in Remote Sensing Images [15.12889076965307]
YOLOv7 one-stage detector is subjected to a novel meta-learning training framework.
This transformation allows the detector to adeptly address FSOD tasks while capitalizing on its inherent advantage of lightweight.
To validate the effectiveness of our proposed detector, we conducted performance comparisons with current state-of-the-art detectors.
arXiv Detail & Related papers (2024-04-29T04:56:52Z) - Small Object Detection via Coarse-to-fine Proposal Generation and
Imitation Learning [52.06176253457522]
We propose a two-stage framework tailored for small object detection based on the Coarse-to-fine pipeline and Feature Imitation learning.
CFINet achieves state-of-the-art performance on the large-scale small object detection benchmarks, SODA-D and SODA-A.
arXiv Detail & Related papers (2023-08-18T13:13:09Z) - Scale-Equivalent Distillation for Semi-Supervised Object Detection [57.59525453301374]
Recent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals.
We analyze the challenges these methods meet with the empirical experiment results.
We introduce a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance.
arXiv Detail & Related papers (2022-03-23T07:33:37Z) - Benchmarking Deep Models for Salient Object Detection [67.07247772280212]
We construct a general SALient Object Detection (SALOD) benchmark to conduct a comprehensive comparison among several representative SOD methods.
In the above experiments, we find that existing loss functions usually specialized in some metrics but reported inferior results on the others.
We propose a novel Edge-Aware (EA) loss that promotes deep networks to learn more discriminative features by integrating both pixel- and image-level supervision signals.
arXiv Detail & Related papers (2022-02-07T03:43:16Z) - Activation to Saliency: Forming High-Quality Labels for Unsupervised
Salient Object Detection [54.92703325989853]
We propose a two-stage Activation-to-Saliency (A2S) framework that effectively generates high-quality saliency cues.
No human annotations are involved in our framework during the whole training process.
Our framework reports significant performance compared with existing USOD methods.
arXiv Detail & Related papers (2021-12-07T11:54:06Z) - Multi-Scale Positive Sample Refinement for Few-Shot Object Detection [61.60255654558682]
Few-shot object detection (FSOD) helps detectors adapt to unseen classes with few training instances.
We propose a Multi-scale Positive Sample Refinement (MPSR) approach to enrich object scales in FSOD.
MPSR generates multi-scale positive samples as object pyramids and refines the prediction at various scales.
arXiv Detail & Related papers (2020-07-18T09:48:29Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.