Partial Is Better Than All: Revisiting Fine-tuning Strategy for Few-shot
Learning
- URL: http://arxiv.org/abs/2102.03983v1
- Date: Mon, 8 Feb 2021 03:27:05 GMT
- Title: Partial Is Better Than All: Revisiting Fine-tuning Strategy for Few-shot
Learning
- Authors: Zhiqiang Shen and Zechun Liu and Jie Qin and Marios Savvides and
Kwang-Ting Cheng
- Abstract summary: A common practice is to train a model on the base set first and then transfer to novel classes through fine-tuning.
We propose to transfer partial knowledge by freezing or fine-tuning particular layer(s) in the base model.
We conduct extensive experiments on CUB and mini-ImageNet to demonstrate the effectiveness of our proposed method.
- Score: 76.98364915566292
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The goal of few-shot learning is to learn a classifier that can recognize
unseen classes from limited support data with labels. A common practice for
this task is to train a model on the base set first and then transfer to novel
classes through fine-tuning (Here fine-tuning procedure is defined as
transferring knowledge from base to novel data, i.e. learning to transfer in
few-shot scenario.) or meta-learning. However, as the base classes have no
overlap to the novel set, simply transferring whole knowledge from base data is
not an optimal solution since some knowledge in the base model may be biased or
even harmful to the novel class. In this paper, we propose to transfer partial
knowledge by freezing or fine-tuning particular layer(s) in the base model.
Specifically, layers will be imposed different learning rates if they are
chosen to be fine-tuned, to control the extent of preserved transferability. To
determine which layers to be recast and what values of learning rates for them,
we introduce an evolutionary search based method that is efficient to
simultaneously locate the target layers and determine their individual learning
rates. We conduct extensive experiments on CUB and mini-ImageNet to demonstrate
the effectiveness of our proposed method. It achieves the state-of-the-art
performance on both meta-learning and non-meta based frameworks. Furthermore,
we extend our method to the conventional pre-training + fine-tuning paradigm
and obtain consistent improvement.
Related papers
- Enhancing Visual Continual Learning with Language-Guided Supervision [76.38481740848434]
Continual learning aims to empower models to learn new tasks without forgetting previously acquired knowledge.
We argue that the scarce semantic information conveyed by the one-hot labels hampers the effective knowledge transfer across tasks.
Specifically, we use PLMs to generate semantic targets for each class, which are frozen and serve as supervision signals.
arXiv Detail & Related papers (2024-03-24T12:41:58Z) - Adaptive Distribution Calibration for Few-Shot Learning with
Hierarchical Optimal Transport [78.9167477093745]
We propose a novel distribution calibration method by learning the adaptive weight matrix between novel samples and base classes.
Experimental results on standard benchmarks demonstrate that our proposed plug-and-play model outperforms competing approaches.
arXiv Detail & Related papers (2022-10-09T02:32:57Z) - Class-Incremental Learning with Strong Pre-trained Models [97.84755144148535]
Class-incremental learning (CIL) has been widely studied under the setting of starting from a small number of classes (base classes)
We explore an understudied real-world setting of CIL that starts with a strong model pre-trained on a large number of base classes.
Our proposed method is robust and generalizes to all analyzed CIL settings.
arXiv Detail & Related papers (2022-04-07T17:58:07Z) - Bridging Non Co-occurrence with Unlabeled In-the-wild Data for
Incremental Object Detection [56.22467011292147]
Several incremental learning methods are proposed to mitigate catastrophic forgetting for object detection.
Despite the effectiveness, these methods require co-occurrence of the unlabeled base classes in the training data of the novel classes.
We propose the use of unlabeled in-the-wild data to bridge the non-occurrence caused by the missing base classes during the training of additional novel classes.
arXiv Detail & Related papers (2021-10-28T10:57:25Z) - Class-incremental Learning with Rectified Feature-Graph Preservation [24.098892115785066]
A central theme of this paper is to learn new classes that arrive in sequential phases over time.
We propose a weighted-Euclidean regularization for old knowledge preservation.
We show how it can work with binary cross-entropy to increase class separation for effective learning of new classes.
arXiv Detail & Related papers (2020-12-15T07:26:04Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - A Primal-Dual Subgradient Approachfor Fair Meta Learning [23.65344558042896]
Few shot meta-learning is well-known with its fast-adapted capability and accuracy generalization onto unseen tasks.
We propose a Primal-Dual Fair Meta-learning framework, namely PDFM, which learns to train fair machine learning models using only a few examples.
arXiv Detail & Related papers (2020-09-26T19:47:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.