Leveraging Normalization Layer in Adapters With Progressive Learning and
Adaptive Distillation for Cross-Domain Few-Shot Learning
- URL: http://arxiv.org/abs/2312.11260v1
- Date: Mon, 18 Dec 2023 15:02:14 GMT
- Title: Leveraging Normalization Layer in Adapters With Progressive Learning and
Adaptive Distillation for Cross-Domain Few-Shot Learning
- Authors: Yongjin Yang, Taehyeon Kim, Se-Young Yun
- Abstract summary: Cross-domain few-shot learning presents a formidable challenge, as models must be trained on base classes and tested on novel classes from various domains with only a few samples at hand.
We introduce a novel generic framework that leverages normalization layer in adapters with Progressive Learning and Adaptive Distillation (ProLAD)
We deploy two strategies: a progressive training of the two adapters and an adaptive distillation technique derived from features determined by the model solely with the adapter devoid of a normalization layer.
- Score: 27.757318834190443
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Cross-domain few-shot learning presents a formidable challenge, as models
must be trained on base classes and then tested on novel classes from various
domains with only a few samples at hand. While prior approaches have primarily
focused on parameter-efficient methods of using adapters, they often overlook
two critical issues: shifts in batch statistics and noisy sample statistics
arising from domain discrepancy variations. In this paper, we introduce a novel
generic framework that leverages normalization layer in adapters with
Progressive Learning and Adaptive Distillation (ProLAD), marking two principal
contributions. First, our methodology utilizes two separate adapters: one
devoid of a normalization layer, which is more effective for similar domains,
and another embedded with a normalization layer, designed to leverage the batch
statistics of the target domain, thus proving effective for dissimilar domains.
Second, to address the pitfalls of noisy statistics, we deploy two strategies:
a progressive training of the two adapters and an adaptive distillation
technique derived from features determined by the model solely with the adapter
devoid of a normalization layer. Through this adaptive distillation, our
approach functions as a modulator, controlling the primary adapter for
adaptation, based on each domain. Evaluations on standard cross-domain few-shot
learning benchmarks confirm that our technique outperforms existing
state-of-the-art methodologies.
Related papers
- Cross-Domain Few-Shot Learning via Adaptive Transformer Networks [16.289485655725013]
This paper proposes an adaptive transformer network (ADAPTER) for cross-domain few-shot learning.
ADAPTER is built upon the idea of bidirectional cross-attention to learn transferable features between the two domains.
arXiv Detail & Related papers (2024-01-25T07:05:42Z) - Adaptive Parametric Prototype Learning for Cross-Domain Few-Shot
Classification [23.82751179819225]
We develop a novel Adaptive Parametric Prototype Learning (APPL) method under the meta-learning convention for cross-domain few-shot classification.
APPL yields superior performance than many state-of-the-art cross-domain few-shot learning methods.
arXiv Detail & Related papers (2023-09-04T03:58:50Z) - Taxonomy Adaptive Cross-Domain Adaptation in Medical Imaging via
Optimization Trajectory Distillation [73.83178465971552]
The success of automated medical image analysis depends on large-scale and expert-annotated training sets.
Unsupervised domain adaptation (UDA) has been raised as a promising approach to alleviate the burden of labeled data collection.
We propose optimization trajectory distillation, a unified approach to address the two technical challenges from a new perspective.
arXiv Detail & Related papers (2023-07-27T08:58:05Z) - NormAUG: Normalization-guided Augmentation for Domain Generalization [60.159546669021346]
We propose a simple yet effective method called NormAUG (Normalization-guided Augmentation) for deep learning.
Our method introduces diverse information at the feature level and improves the generalization of the main path.
In the test stage, we leverage an ensemble strategy to combine the predictions from the auxiliary path of our model, further boosting performance.
arXiv Detail & Related papers (2023-07-25T13:35:45Z) - Learning Instance-Specific Adaptation for Cross-Domain Segmentation [79.61787982393238]
We propose a test-time adaptation method for cross-domain image segmentation.
Given a new unseen instance at test time, we adapt a pre-trained model by conducting instance-specific BatchNorm calibration.
arXiv Detail & Related papers (2022-03-30T17:59:45Z) - Style Mixing and Patchwise Prototypical Matching for One-Shot
Unsupervised Domain Adaptive Semantic Segmentation [21.01132797297286]
In one-shot unsupervised domain adaptation, segmentors only see one unlabeled target image during training.
We propose a new OSUDA method that can effectively relieve such computational burden.
Our method achieves new state-of-the-art performance on two commonly used benchmarks for domain adaptive semantic segmentation.
arXiv Detail & Related papers (2021-12-09T02:47:46Z) - Adaptively-Accumulated Knowledge Transfer for Partial Domain Adaptation [66.74638960925854]
Partial domain adaptation (PDA) deals with a realistic and challenging problem when the source domain label space substitutes the target domain.
We propose an Adaptively-Accumulated Knowledge Transfer framework (A$2$KT) to align the relevant categories across two domains.
arXiv Detail & Related papers (2020-08-27T00:53:43Z) - Class-Incremental Domain Adaptation [56.72064953133832]
We introduce a practical Domain Adaptation (DA) paradigm called Class-Incremental Domain Adaptation (CIDA)
Existing DA methods tackle domain-shift but are unsuitable for learning novel target-domain classes.
Our approach yields superior performance as compared to both DA and CI methods in the CIDA paradigm.
arXiv Detail & Related papers (2020-08-04T07:55:03Z) - Towards Fair Cross-Domain Adaptation via Generative Learning [50.76694500782927]
Domain Adaptation (DA) targets at adapting a model trained over the well-labeled source domain to the unlabeled target domain lying in different distributions.
We develop a novel Generative Few-shot Cross-domain Adaptation (GFCA) algorithm for fair cross-domain classification.
arXiv Detail & Related papers (2020-03-04T23:25:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.