Generalized Robust Test-Time Adaptation in Continuous Dynamic Scenarios
- URL: http://arxiv.org/abs/2310.04714v1
- Date: Sat, 7 Oct 2023 07:13:49 GMT
- Title: Generalized Robust Test-Time Adaptation in Continuous Dynamic Scenarios
- Authors: Shuang Li, Longhui Yuan, Binhui Xie and Tao Yang
- Abstract summary: Test-time adaptation (TTA) adapts pre-trained models to test distributions during the inference phase exclusively employing unlabeled test data streams.
We propose a Generalized Robust Test-Time Adaptation (GRoTTA) method to effectively address the difficult problem.
- Score: 18.527640606971563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Test-time adaptation (TTA) adapts the pre-trained models to test
distributions during the inference phase exclusively employing unlabeled test
data streams, which holds great value for the deployment of models in
real-world applications. Numerous studies have achieved promising performance
on simplistic test streams, characterized by independently and uniformly
sampled test data originating from a fixed target data distribution. However,
these methods frequently prove ineffective in practical scenarios, where both
continual covariate shift and continual label shift occur simultaneously, i.e.,
data and label distributions change concurrently and continually over time. In
this study, a more challenging Practical Test-Time Adaptation (PTTA) setup is
introduced, which takes into account the concurrent presence of continual
covariate shift and continual label shift, and we propose a Generalized Robust
Test-Time Adaptation (GRoTTA) method to effectively address the difficult
problem. We start by steadily adapting the model through Robust Parameter
Adaptation to make balanced predictions for test samples. To be specific,
firstly, the effects of continual label shift are eliminated by enforcing the
model to learn from a uniform label distribution and introducing recalibration
of batch normalization to ensure stability. Secondly, the continual covariate
shift is alleviated by employing a source knowledge regularization with the
teacher-student model to update parameters. Considering the potential
information in the test stream, we further refine the balanced predictions by
Bias-Guided Output Adaptation, which exploits latent structure in the feature
space and is adaptive to the imbalanced label distribution. Extensive
experiments demonstrate GRoTTA outperforms the existing competitors by a large
margin under PTTA setting, rendering it highly conducive for adoption in
real-world applications.
Related papers
- DOTA: Distributional Test-Time Adaptation of Vision-Language Models [52.98590762456236]
Training-free test-time dynamic adapter (TDA) is a promising approach to address this issue.
We propose a simple yet effective method for DistributiOnal Test-time Adaptation (Dota)
Dota continually estimates the distributions of test samples, allowing the model to continually adapt to the deployment environment.
arXiv Detail & Related papers (2024-09-28T15:03:28Z) - Uncertainty-Calibrated Test-Time Model Adaptation without Forgetting [55.17761802332469]
Test-time adaptation (TTA) seeks to tackle potential distribution shifts between training and test data by adapting a given model w.r.t. any test sample.
Prior methods perform backpropagation for each test sample, resulting in unbearable optimization costs to many applications.
We propose an Efficient Anti-Forgetting Test-Time Adaptation (EATA) method which develops an active sample selection criterion to identify reliable and non-redundant samples.
arXiv Detail & Related papers (2024-03-18T05:49:45Z) - Robust Test-Time Adaptation in Dynamic Scenarios [9.475271284789969]
Test-time adaptation (TTA) intends to adapt the pretrained model to test distributions with only unlabeled test data streams.
We elaborate a Robust Test-Time Adaptation (RoTTA) method against the complex data stream in PTTA.
Our method is easy to implement, making it a good choice for rapid deployment.
arXiv Detail & Related papers (2023-03-24T10:19:14Z) - DELTA: degradation-free fully test-time adaptation [59.74287982885375]
We find that two unfavorable defects are concealed in the prevalent adaptation methodologies like test-time batch normalization (BN) and self-learning.
First, we reveal that the normalization statistics in test-time BN are completely affected by the currently received test samples, resulting in inaccurate estimates.
Second, we show that during test-time adaptation, the parameter update is biased towards some dominant classes.
arXiv Detail & Related papers (2023-01-30T15:54:00Z) - Rethinking Precision of Pseudo Label: Test-Time Adaptation via
Complementary Learning [10.396596055773012]
We propose a novel complementary learning approach to enhance test-time adaptation.
In test-time adaptation tasks, information from the source domain is typically unavailable.
We highlight that the risk function of complementary labels agrees with their Vanilla loss formula.
arXiv Detail & Related papers (2023-01-15T03:36:33Z) - Robust Continual Test-time Adaptation: Instance-aware BN and
Prediction-balanced Memory [58.72445309519892]
We present a new test-time adaptation scheme that is robust against non-i.i.d. test data streams.
Our novelty is mainly two-fold: (a) Instance-Aware Batch Normalization (IABN) that corrects normalization for out-of-distribution samples, and (b) Prediction-balanced Reservoir Sampling (PBRS) that simulates i.i.d. data stream from non-i.i.d. stream in a class-balanced manner.
arXiv Detail & Related papers (2022-08-10T03:05:46Z) - Efficient Test-Time Model Adaptation without Forgetting [60.36499845014649]
Test-time adaptation seeks to tackle potential distribution shifts between training and testing data.
We propose an active sample selection criterion to identify reliable and non-redundant samples.
We also introduce a Fisher regularizer to constrain important model parameters from drastic changes.
arXiv Detail & Related papers (2022-04-06T06:39:40Z) - Training on Test Data with Bayesian Adaptation for Covariate Shift [96.3250517412545]
Deep neural networks often make inaccurate predictions with unreliable uncertainty estimates.
We derive a Bayesian model that provides for a well-defined relationship between unlabeled inputs under distributional shift and model parameters.
We show that our method improves both accuracy and uncertainty estimation.
arXiv Detail & Related papers (2021-09-27T01:09:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.