Visual Prompt Tuning for Test-time Domain Adaptation
- URL: http://arxiv.org/abs/2210.04831v1
- Date: Mon, 10 Oct 2022 16:45:13 GMT
- Title: Visual Prompt Tuning for Test-time Domain Adaptation
- Authors: Yunhe Gao, Xingjian Shi, Yi Zhu, Hao Wang, Zhiqiang Tang, Xiong Zhou,
Mu Li, Dimitris N. Metaxas
- Abstract summary: We propose a simple recipe called data-efficient prompt tuning (DePT) with two key ingredients.
We find such parameter-efficient finetuning can efficiently adapt the model representation to the target domain without overfitting to the noise in the learning objective.
With much fewer parameters, DePT demonstrates not only state-of-the-art performance on major adaptation benchmarks, but also superior data efficiency.
- Score: 48.16620171809511
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Models should have the ability to adapt to unseen data during test-time to
avoid performance drop caused by inevitable distribution shifts in real-world
deployment scenarios. In this work, we tackle the practical yet challenging
test-time adaptation (TTA) problem, where a model adapts to the target domain
without accessing the source data. We propose a simple recipe called
data-efficient prompt tuning (DePT) with two key ingredients. First, DePT plugs
visual prompts into the vision Transformer and only tunes these
source-initialized prompts during adaptation. We find such parameter-efficient
finetuning can efficiently adapt the model representation to the target domain
without overfitting to the noise in the learning objective. Second, DePT
bootstraps the source representation to the target domain by memory bank-based
online pseudo labeling. A hierarchical self-supervised regularization specially
designed for prompts is jointly optimized to alleviate error accumulation
during self-training. With much fewer tunable parameters, DePT demonstrates not
only state-of-the-art performance on major adaptation benchmarks, but also
superior data efficiency, i.e., adaptation with only 1\% or 10\% data without
much performance degradation compared to 100\% data. In addition, DePT is also
versatile to be extended to online or multi-source TTA settings.
Related papers
- Visual Fourier Prompt Tuning [63.66866445034855]
We propose the Visual Fourier Prompt Tuning (VFPT) method as a general and effective solution for adapting large-scale transformer-based models.
Our approach incorporates the Fast Fourier Transform into prompt embeddings and harmoniously considers both spatial and frequency domain information.
Our results demonstrate that our approach outperforms current state-of-the-art baselines on two benchmarks.
arXiv Detail & Related papers (2024-11-02T18:18:35Z) - Parameter-Efficient Fine-Tuning With Adapters [5.948206235442328]
This research introduces a novel adaptation method utilizing the UniPELT framework as a base.
Our method employs adapters, which enable efficient transfer of pretrained models to new tasks with minimal retraining of the base model parameters.
arXiv Detail & Related papers (2024-05-09T01:40:38Z) - REALM: Robust Entropy Adaptive Loss Minimization for Improved
Single-Sample Test-Time Adaptation [5.749155230209001]
Fully-test-time adaptation (F-TTA) can mitigate performance loss due to distribution shifts between train and test data.
We present a general framework for improving robustness of F-TTA to noisy samples, inspired by self-paced learning and robust loss functions.
arXiv Detail & Related papers (2023-09-07T18:44:58Z) - Approximated Prompt Tuning for Vision-Language Pre-trained Models [54.326232586461614]
In vision-language pre-trained models, prompt tuning often requires a large number of learnable tokens to bridge the gap between the pre-training and downstream tasks.
We propose a novel Approximated Prompt Tuning (APT) approach towards efficient VL transfer learning.
arXiv Detail & Related papers (2023-06-27T05:43:47Z) - Sensitivity-Aware Visual Parameter-Efficient Fine-Tuning [91.5113227694443]
We propose a novel visual.
sensuous-aware fine-Tuning (SPT) scheme.
SPT allocates trainable parameters to task-specific important positions.
Experiments on a wide range of downstream recognition tasks show that our SPT is complementary to the existing PEFT methods.
arXiv Detail & Related papers (2023-03-15T12:34:24Z) - Improving the Sample Efficiency of Prompt Tuning with Domain Adaptation [15.388175691903252]
We propose doMain Adaptation (OPTIMA) to regularizes the decision boundary to be smooth around regions where source and target data distributions are similar.
OPTIMA significantly enhances the transferability and sample-efficiency of prompt tuning compared to strong baselines.
arXiv Detail & Related papers (2022-10-06T14:44:21Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Listen, Adapt, Better WER: Source-free Single-utterance Test-time
Adaptation for Automatic Speech Recognition [65.84978547406753]
Test-time Adaptation aims to adapt the model trained on source domains to yield better predictions for test samples.
Single-Utterance Test-time Adaptation (SUTA) is the first TTA study in speech area to our best knowledge.
arXiv Detail & Related papers (2022-03-27T06:38:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.