LPT: Long-tailed Prompt Tuning for Image Classification
- URL: http://arxiv.org/abs/2210.01033v2
- Date: Tue, 28 Mar 2023 10:16:03 GMT
- Title: LPT: Long-tailed Prompt Tuning for Image Classification
- Authors: Bowen Dong, Pan Zhou, Shuicheng Yan, Wangmeng Zuo
- Abstract summary: We introduce several trainable prompts into a frozen pretrained model to adapt it to long-tailed data.
In phase 1, we train the shared prompt via supervised prompt tuning to adapt a pretrained model to the desired long-tailed domain.
In phase 2, we use the learnt shared prompt as query to select a small best matched set for a group of similar samples.
- Score: 178.52948452353834
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: For long-tailed classification, most works often pretrain a big model on a
large-scale dataset, and then fine-tune the whole model for adapting to
long-tailed data. Though promising, fine-tuning the whole pretrained model
tends to suffer from high cost in computation and deployment of different
models for different tasks, as well as weakened generalization ability for
overfitting to certain features of long-tailed data. To alleviate these issues,
we propose an effective Long-tailed Prompt Tuning method for long-tailed
classification. LPT introduces several trainable prompts into a frozen
pretrained model to adapt it to long-tailed data. For better effectiveness, we
divide prompts into two groups: 1) a shared prompt for the whole long-tailed
dataset to learn general features and to adapt a pretrained model into target
domain; and 2) group-specific prompts to gather group-specific features for the
samples which have similar features and also to empower the pretrained model
with discrimination ability. Then we design a two-phase training paradigm to
learn these prompts. In phase 1, we train the shared prompt via supervised
prompt tuning to adapt a pretrained model to the desired long-tailed domain. In
phase 2, we use the learnt shared prompt as query to select a small best
matched set for a group of similar samples from the group-specific prompt set
to dig the common features of these similar samples, then optimize these
prompts with dual sampling strategy and asymmetric GCL loss. By only
fine-tuning a few prompts while fixing the pretrained model, LPT can reduce
training and deployment cost by storing a few prompts, and enjoys a strong
generalization ability of the pretrained model. Experiments show that on
various long-tailed benchmarks, with only ~1.1% extra parameters, LPT achieves
comparable performance than previous whole model fine-tuning methods, and is
more robust to domain-shift.
Related papers
- Adapt-$\infty$: Scalable Lifelong Multimodal Instruction Tuning via Dynamic Data Selection [89.42023974249122]
Adapt-$infty$ is a new multi-way and adaptive data selection approach for Lifelong Instruction Tuning.
We construct pseudo-skill clusters by grouping gradient-based sample vectors.
We select the best-performing data selector for each skill cluster from a pool of selector experts.
arXiv Detail & Related papers (2024-10-14T15:48:09Z) - LPT++: Efficient Training on Mixture of Long-tailed Experts [107.78420448806357]
++ enhances frozen Vision Transformers (ViTs) through the integration of three core components.
The first is a universal long-tailed adaptation module, which aggregates long-tailed prompts and visual adapters to adapt the pretrained model to the target domain.
The second is the mixture of long-tailed experts framework with a mixture-of-experts (MoE) scorer, which adaptively calculates reweighting coefficients for confidence scores from both visual-only and visual-language (VL) model experts to generate more accurate predictions.
arXiv Detail & Related papers (2024-09-17T16:19:11Z) - GistScore: Learning Better Representations for In-Context Example
Selection with Gist Bottlenecks [3.9638110494107095]
In-context Learning (ICL) is the ability of Large Language Models (LLMs) to perform new tasks when conditioned on prompts.
We propose Example Gisting, a novel approach for training example encoders through supervised fine-tuning.
We show that our fine-tuned models get state-of-the-art ICL performance with over 20% absolute gain over off-the-shelf retrievers.
arXiv Detail & Related papers (2023-11-16T06:28:05Z) - RanPAC: Random Projections and Pre-trained Models for Continual Learning [59.07316955610658]
Continual learning (CL) aims to learn different tasks (such as classification) in a non-stationary data stream without forgetting old ones.
We propose a concise and effective approach for CL with pre-trained models.
arXiv Detail & Related papers (2023-07-05T12:49:02Z) - $\Delta$-Patching: A Framework for Rapid Adaptation of Pre-trained
Convolutional Networks without Base Performance Loss [71.46601663956521]
Models pre-trained on large-scale datasets are often fine-tuned to support newer tasks and datasets that arrive over time.
We propose $Delta$-Patching for fine-tuning neural network models in an efficient manner, without the need to store model copies.
Our experiments show that $Delta$-Networks outperform earlier model patching work while only requiring a fraction of parameters to be trained.
arXiv Detail & Related papers (2023-03-26T16:39:44Z) - Federated Adaptive Prompt Tuning for Multi-Domain Collaborative Learning [44.604485649167216]
Federated learning (FL) enables multiple clients to collaboratively train a global model without disclosing their data.
We propose a federated adaptive prompt tuning algorithm, FedAPT, for multi-domain collaborative image classification.
arXiv Detail & Related papers (2022-11-15T03:10:05Z) - Test-Time Prompt Tuning for Zero-Shot Generalization in Vision-Language
Models [107.05966685291067]
We propose test-time prompt tuning (TPT) to learn adaptive prompts on the fly with a single test sample.
TPT improves the zero-shot top-1 accuracy of CLIP by 3.6% on average.
In evaluating cross-dataset generalization with unseen categories, TPT performs on par with the state-of-the-art approaches that use additional training data.
arXiv Detail & Related papers (2022-09-15T17:55:11Z) - Deep Ensembles for Low-Data Transfer Learning [21.578470914935938]
We study different ways of creating ensembles from pre-trained models.
We show that the nature of pre-training itself is a performant source of diversity.
We propose a practical algorithm that efficiently identifies a subset of pre-trained models for any downstream dataset.
arXiv Detail & Related papers (2020-10-14T07:59:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.