Label Shift Adapter for Test-Time Adaptation under Covariate and Label
Shifts
- URL: http://arxiv.org/abs/2308.08810v1
- Date: Thu, 17 Aug 2023 06:37:37 GMT
- Title: Label Shift Adapter for Test-Time Adaptation under Covariate and Label
Shifts
- Authors: Sunghyun Park, Seunghan Yang, Jaegul Choo, Sungrack Yun
- Abstract summary: Test-time adaptation (TTA) aims to adapt a pre-trained model to the target domain in a batch-by-batch manner during inference.
Most previous TTA approaches assume that both source and target domain datasets have balanced label distribution.
We propose a novel label shift adapter that can be incorporated into existing TTA approaches to deal with label shifts effectively.
- Score: 48.83127071288469
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Test-time adaptation (TTA) aims to adapt a pre-trained model to the target
domain in a batch-by-batch manner during inference. While label distributions
often exhibit imbalances in real-world scenarios, most previous TTA approaches
typically assume that both source and target domain datasets have balanced
label distribution. Due to the fact that certain classes appear more frequently
in certain domains (e.g., buildings in cities, trees in forests), it is natural
that the label distribution shifts as the domain changes. However, we discover
that the majority of existing TTA methods fail to address the coexistence of
covariate and label shifts. To tackle this challenge, we propose a novel label
shift adapter that can be incorporated into existing TTA approaches to deal
with label shifts during the TTA process effectively. Specifically, we estimate
the label distribution of the target domain to feed it into the label shift
adapter. Subsequently, the label shift adapter produces optimal parameters for
the target label distribution. By predicting only the parameters for a part of
the pre-trained source model, our approach is computationally efficient and can
be easily applied, regardless of the model architectures. Through extensive
experiments, we demonstrate that integrating our strategy with TTA approaches
leads to substantial performance improvements under the joint presence of label
and covariate shifts.
Related papers
- Channel-Selective Normalization for Label-Shift Robust Test-Time Adaptation [16.657929958093824]
Test-time adaptation is an approach to adjust models to a new data distribution during inference.
Test-time batch normalization is a simple and popular method that achieved compelling performance on domain shift benchmarks.
We propose to tackle this challenge by only selectively adapting channels in a deep network, minimizing drastic adaptation that is sensitive to label shifts.
arXiv Detail & Related papers (2024-02-07T15:41:01Z) - ELSA: Efficient Label Shift Adaptation through the Lens of
Semiparametric Models [2.73424570009809]
Under the label shift context, the marginal distribution of the label varies across the training and testing datasets.
We propose a moment-matching framework for adapting the label shift based on the geometry of the influence function.
Under such a framework, we propose a novel method named underlineEfficient underlineLabel underlineShift underlineAdaptation (ELSA)
arXiv Detail & Related papers (2023-05-30T15:31:44Z) - RLSbench: Domain Adaptation Under Relaxed Label Shift [39.845383643588356]
We introduce RLSbench, a large-scale benchmark for relaxed label shift.
We assess 13 popular domain adaptation methods, demonstrating more widespread failures under label proportion shifts than were previously known.
We develop an effective two-step meta-algorithm that is compatible with most domain adaptations.
arXiv Detail & Related papers (2023-02-06T18:57:14Z) - Domain Adaptation under Open Set Label Shift [39.424134505152544]
We introduce the problem of domain adaptation under Open Set Label Shift (OSLS)
OSLS subsumes domain adaptation under label shift and Positive-Unlabeled (PU) learning.
We propose practical methods for both tasks that leverage black-box predictors.
arXiv Detail & Related papers (2022-07-26T17:09:48Z) - Instance-Dependent Partial Label Learning [69.49681837908511]
Partial label learning is a typical weakly supervised learning problem.
Most existing approaches assume that the incorrect labels in each training example are randomly picked as the candidate labels.
In this paper, we consider instance-dependent and assume that each example is associated with a latent label distribution constituted by the real number of each label.
arXiv Detail & Related papers (2021-10-25T12:50:26Z) - Semi-Supervised Domain Adaptation with Prototypical Alignment and
Consistency Learning [86.6929930921905]
This paper studies how much it can help address domain shifts if we further have a few target samples labeled.
To explore the full potential of landmarks, we incorporate a prototypical alignment (PA) module which calculates a target prototype for each class from the landmarks.
Specifically, we severely perturb the labeled images, making PA non-trivial to achieve and thus promoting model generalizability.
arXiv Detail & Related papers (2021-04-19T08:46:08Z) - Cycle Self-Training for Domain Adaptation [85.14659717421533]
Cycle Self-Training (CST) is a principled self-training algorithm that enforces pseudo-labels to generalize across domains.
CST recovers target ground truth, while both invariant feature learning and vanilla self-training fail.
Empirical results indicate that CST significantly improves over prior state-of-the-arts in standard UDA benchmarks.
arXiv Detail & Related papers (2021-03-05T10:04:25Z) - Coping with Label Shift via Distributionally Robust Optimisation [72.80971421083937]
We propose a model that minimises an objective based on distributionally robust optimisation (DRO)
We then design and analyse a gradient descent-proximal mirror ascent algorithm tailored for large-scale problems to optimise the proposed objective.
arXiv Detail & Related papers (2020-10-23T08:33:04Z) - Few-shot Slot Tagging with Collapsed Dependency Transfer and
Label-enhanced Task-adaptive Projection Network [61.94394163309688]
We propose a Label-enhanced Task-Adaptive Projection Network (L-TapNet) based on the state-of-the-art few-shot classification model -- TapNet.
Experimental results show that our model significantly outperforms the strongest few-shot learning baseline by 14.64 F1 scores in the one-shot setting.
arXiv Detail & Related papers (2020-06-10T07:50:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.