Few-Shot Adaptation of Pre-Trained Networks for Domain Shift
- URL: http://arxiv.org/abs/2205.15234v1
- Date: Mon, 30 May 2022 16:49:59 GMT
- Title: Few-Shot Adaptation of Pre-Trained Networks for Domain Shift
- Authors: Wenyu Zhang, Li Shen, Wanyue Zhang, Chuan-Sheng Foo
- Abstract summary: Deep networks are prone to performance degradation when there is a domain shift between the source (training) data and target (test) data.
Recent test-time adaptation methods update batch normalization layers of pre-trained source models deployed in new target environments with streaming data to mitigate such performance degradation.
We propose a framework for few-shot domain adaptation to address the practical challenges of data-efficient adaptation.
- Score: 17.123505029637055
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep networks are prone to performance degradation when there is a domain
shift between the source (training) data and target (test) data. Recent
test-time adaptation methods update batch normalization layers of pre-trained
source models deployed in new target environments with streaming data to
mitigate such performance degradation. Although such methods can adapt
on-the-fly without first collecting a large target domain dataset, their
performance is dependent on streaming conditions such as mini-batch size and
class-distribution, which can be unpredictable in practice. In this work, we
propose a framework for few-shot domain adaptation to address the practical
challenges of data-efficient adaptation. Specifically, we propose a constrained
optimization of feature normalization statistics in pre-trained source models
supervised by a small support set from the target domain. Our method is easy to
implement and improves source model performance with as few as one sample per
class for classification tasks. Extensive experiments on 5 cross-domain
classification and 4 semantic segmentation datasets show that our method
achieves more accurate and reliable performance than test-time adaptation,
while not being constrained by streaming conditions.
Related papers
- Meta-TTT: A Meta-learning Minimax Framework For Test-Time Training [5.9631503543049895]
Test-time domain adaptation is a challenging task that aims to adapt a pre-trained model to limited, unlabeled target data during inference.
This paper introduces a meta-learning minimax framework for test-time training on batch normalization layers.
arXiv Detail & Related papers (2024-10-02T16:16:05Z) - Efficient Transferability Assessment for Selection of Pre-trained Detectors [63.21514888618542]
This paper studies the efficient transferability assessment of pre-trained object detectors.
We build up a detector transferability benchmark which contains a large and diverse zoo of pre-trained detectors.
Experimental results demonstrate that our method outperforms other state-of-the-art approaches in assessing transferability.
arXiv Detail & Related papers (2024-03-14T14:23:23Z) - Adaptive Training Distributions with Scalable Online Bilevel
Optimization [26.029033134519604]
Large neural networks pretrained on web-scale corpora are central to modern machine learning.
This work considers modifying the pretraining distribution in the case where one has a small sample of data reflecting the targeted test conditions.
We propose an algorithm motivated by a recent formulation of this setting as an online, bilevel optimization problem.
arXiv Detail & Related papers (2023-11-20T18:01:29Z) - Informative Data Mining for One-Shot Cross-Domain Semantic Segmentation [84.82153655786183]
We propose a novel framework called Informative Data Mining (IDM) to enable efficient one-shot domain adaptation for semantic segmentation.
IDM provides an uncertainty-based selection criterion to identify the most informative samples, which facilitates quick adaptation and reduces redundant training.
Our approach outperforms existing methods and achieves a new state-of-the-art one-shot performance of 56.7%/55.4% on the GTA5/SYNTHIA to Cityscapes adaptation tasks.
arXiv Detail & Related papers (2023-09-25T15:56:01Z) - Addressing Distribution Shift at Test Time in Pre-trained Language
Models [3.655021726150369]
State-of-the-art pre-trained language models (PLMs) outperform other models when applied to the majority of language processing tasks.
PLMs have been found to degrade in performance under distribution shift.
We present an approach that improves the performance of PLMs at test-time under distribution shift.
arXiv Detail & Related papers (2022-12-05T16:04:54Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Continual Test-Time Domain Adaptation [94.51284735268597]
Test-time domain adaptation aims to adapt a source pre-trained model to a target domain without using any source data.
CoTTA is easy to implement and can be readily incorporated in off-the-shelf pre-trained models.
arXiv Detail & Related papers (2022-03-25T11:42:02Z) - Adaptive Risk Minimization: Learning to Adapt to Domain Shift [109.87561509436016]
A fundamental assumption of most machine learning algorithms is that the training and test data are drawn from the same underlying distribution.
In this work, we consider the problem setting of domain generalization, where the training data are structured into domains and there may be multiple test time shifts.
We introduce the framework of adaptive risk minimization (ARM), in which models are directly optimized for effective adaptation to shift by learning to adapt on the training domains.
arXiv Detail & Related papers (2020-07-06T17:59:30Z) - Don't Stop Pretraining: Adapt Language Models to Domains and Tasks [81.99843216550306]
We present a study across four domains (biomedical and computer science publications, news, and reviews) and eight classification tasks.
A second phase of pretraining in-domain (domain-adaptive pretraining) leads to performance gains.
Adapting to the task's unlabeled data (task-adaptive pretraining) improves performance even after domain-adaptive pretraining.
arXiv Detail & Related papers (2020-04-23T04:21:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.