Efficient Few-Shot Object Detection via Knowledge Inheritance
- URL: http://arxiv.org/abs/2203.12224v1
- Date: Wed, 23 Mar 2022 06:24:31 GMT
- Title: Efficient Few-Shot Object Detection via Knowledge Inheritance
- Authors: Ze Yang, Chi Zhang, Ruibo Li and Guosheng Lin
- Abstract summary: Few-shot object detection (FSOD) aims at learning a generic detector that can adapt to unseen tasks with scarce training samples.
We present an efficient pretrain-transfer framework (PTF) baseline with no computational increment.
We also propose an adaptive length re-scaling (ALR) strategy to alleviate the vector length inconsistency between the predicted novel weights and the pretrained base weights.
- Score: 62.36414544915032
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot object detection (FSOD), which aims at learning a generic detector
that can adapt to unseen tasks with scarce training samples, has witnessed
consistent improvement recently. However, most existing methods ignore the
efficiency issues, e.g., high computational complexity and slow adaptation
speed. Notably, efficiency has become an increasingly important evaluation
metric for few-shot techniques due to an emerging trend toward embedded AI. To
this end, we present an efficient pretrain-transfer framework (PTF) baseline
with no computational increment, which achieves comparable results with
previous state-of-the-art (SOTA) methods. Upon this baseline, we devise an
initializer named knowledge inheritance (KI) to reliably initialize the novel
weights for the box classifier, which effectively facilitates the knowledge
transfer process and boosts the adaptation speed. Within the KI initializer, we
propose an adaptive length re-scaling (ALR) strategy to alleviate the vector
length inconsistency between the predicted novel weights and the pretrained
base weights. Finally, our approach not only achieves the SOTA results across
three public benchmarks, i.e., PASCAL VOC, COCO and LVIS, but also exhibits
high efficiency with 1.8-9.0x faster adaptation speed against the other methods
on COCO/LVIS benchmark during few-shot transfer. To our best knowledge, this is
the first work to consider the efficiency problem in FSOD. We hope to motivate
a trend toward powerful yet efficient few-shot technique development. The codes
are publicly available at https://github.com/Ze-Yang/Efficient-FSOD.
Related papers
- VeLoRA: Memory Efficient Training using Rank-1 Sub-Token Projections [35.133698935322634]
Large language models (LLMs) have recently emerged as powerful tools for tackling many language-processing tasks.
We identify and characterise the important components needed for effective model convergence using gradient descent.
This result leads us to a cheap and memory-efficient algorithm for both fine-tuning and pre-training LLMs.
arXiv Detail & Related papers (2024-05-28T09:23:14Z) - Efficient Training for Visual Tracking with Deformable Transformer [0.0]
We present DETRack, a streamlined end-to-end visual object tracking framework.
Our framework utilizes an efficient encoder-decoder structure where the deformable transformer decoder acting as a target head.
For training, we introduce a novel one-to-many label assignment and an auxiliary denoising technique.
arXiv Detail & Related papers (2023-09-06T03:07:43Z) - Boosting Fast Adversarial Training with Learnable Adversarial
Initialization [79.90495058040537]
Adrial training (AT) has been demonstrated to be effective in improving model robustness by leveraging adversarial examples for training.
To boost training efficiency, fast gradient sign method (FGSM) is adopted in fast AT methods by calculating gradient only once.
arXiv Detail & Related papers (2021-10-11T05:37:00Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - Prior Guided Feature Enrichment Network for Few-Shot Segmentation [64.91560451900125]
State-of-the-art semantic segmentation methods require sufficient labeled data to achieve good results.
Few-shot segmentation is proposed to tackle this problem by learning a model that quickly adapts to new classes with a few labeled support samples.
Theses frameworks still face the challenge of generalization ability reduction on unseen classes due to inappropriate use of high-level semantic information.
arXiv Detail & Related papers (2020-08-04T10:41:32Z) - AdamP: Slowing Down the Slowdown for Momentum Optimizers on
Scale-invariant Weights [53.8489656709356]
Normalization techniques are a boon for modern deep learning.
It is often overlooked, however, that the additional introduction of momentum results in a rapid reduction in effective step sizes for scale-invariant weights.
In this paper, we verify that the widely-adopted combination of the two ingredients lead to the premature decay of effective step sizes and sub-optimal model performances.
arXiv Detail & Related papers (2020-06-15T08:35:15Z) - MC-BERT: Efficient Language Pre-Training via a Meta Controller [96.68140474547602]
Large-scale pre-training is computationally expensive.
ELECTRA, an early attempt to accelerate pre-training, trains a discriminative model that predicts whether each input token was replaced by a generator.
We propose a novel meta-learning framework, MC-BERT, to achieve better efficiency and effectiveness.
arXiv Detail & Related papers (2020-06-10T09:22:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.