A Probabilistic Framework for Lifelong Test-Time Adaptation
- URL: http://arxiv.org/abs/2212.09713v2
- Date: Tue, 4 Apr 2023 07:52:40 GMT
- Title: A Probabilistic Framework for Lifelong Test-Time Adaptation
- Authors: Dhanajit Brahma and Piyush Rai
- Abstract summary: Test-time adaptation (TTA) is the problem of updating a pre-trained source model at inference time given test input(s) from a different target domain.
We present PETAL (Probabilistic lifElong Test-time Adaptation with seLf-training prior), which solves lifelong TTA using a probabilistic approach.
Our method achieves better results than the current state-of-the-art for online lifelong test-time adaptation across various benchmarks.
- Score: 34.07074915005366
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Test-time adaptation (TTA) is the problem of updating a pre-trained source
model at inference time given test input(s) from a different target domain.
Most existing TTA approaches assume the setting in which the target domain is
stationary, i.e., all the test inputs come from a single target domain.
However, in many practical settings, the test input distribution might exhibit
a lifelong/continual shift over time. Moreover, existing TTA approaches also
lack the ability to provide reliable uncertainty estimates, which is crucial
when distribution shifts occur between the source and target domain. To address
these issues, we present PETAL (Probabilistic lifElong Test-time Adaptation
with seLf-training prior), which solves lifelong TTA using a probabilistic
approach, and naturally results in (1) a student-teacher framework, where the
teacher model is an exponential moving average of the student model, and (2)
regularizing the model updates at inference time using the source model as a
regularizer. To prevent model drift in the lifelong/continual TTA setting, we
also propose a data-driven parameter restoration technique which contributes to
reducing the error accumulation and maintaining the knowledge of recent domains
by restoring only the irrelevant parameters. In terms of predictive error rate
as well as uncertainty based metrics such as Brier score and negative
log-likelihood, our method achieves better results than the current
state-of-the-art for online lifelong test-time adaptation across various
benchmarks, such as CIFAR-10C, CIFAR-100C, ImageNetC, and ImageNet3DCC
datasets. The source code for our approach is accessible at
https://github.com/dhanajitb/petal.
Related papers
- DOTA: Distributional Test-Time Adaptation of Vision-Language Models [52.98590762456236]
Training-free test-time dynamic adapter (TDA) is a promising approach to address this issue.
We propose a simple yet effective method for DistributiOnal Test-time Adaptation (Dota)
Dota continually estimates the distributions of test samples, allowing the model to continually adapt to the deployment environment.
arXiv Detail & Related papers (2024-09-28T15:03:28Z) - ETAGE: Enhanced Test Time Adaptation with Integrated Entropy and Gradient Norms for Robust Model Performance [18.055032898349438]
Test time adaptation (TTA) equips deep learning models to handle unseen test data that deviates from the training distribution.
We introduce ETAGE, a refined TTA method that integrates entropy minimization with gradient norms and PLPD.
Our method prioritizes samples that are less likely to cause instability by combining high entropy with high gradient norms out of adaptation.
arXiv Detail & Related papers (2024-09-14T01:25:52Z) - Persistent Test-time Adaptation in Recurring Testing Scenarios [12.024233973321756]
Current test-time adaptation (TTA) approaches aim to adapt a machine learning model to environments that change continuously.
Yet, it is unclear whether TTA methods can maintain their adaptability over prolonged periods.
We propose persistent TTA (PeTTA) which senses when the model is diverging towards collapse and adjusts the adaptation strategy.
arXiv Detail & Related papers (2023-11-30T02:24:44Z) - DELTA: degradation-free fully test-time adaptation [59.74287982885375]
We find that two unfavorable defects are concealed in the prevalent adaptation methodologies like test-time batch normalization (BN) and self-learning.
First, we reveal that the normalization statistics in test-time BN are completely affected by the currently received test samples, resulting in inaccurate estimates.
Second, we show that during test-time adaptation, the parameter update is biased towards some dominant classes.
arXiv Detail & Related papers (2023-01-30T15:54:00Z) - Robust Mean Teacher for Continual and Gradual Test-Time Adaptation [5.744133015573047]
Gradual test-time adaptation (TTA) considers not only a single domain shift, but a sequence of shifts.
We propose and show that in the setting of TTA, the symmetric cross-entropy is better suited as a consistency loss for mean teachers.
We demonstrate the effectiveness of our proposed method 'robust mean teacher' (RMT) on the continual and gradual corruption benchmarks.
arXiv Detail & Related papers (2022-11-23T16:14:45Z) - TeST: Test-time Self-Training under Distribution Shift [99.68465267994783]
Test-Time Self-Training (TeST) is a technique that takes as input a model trained on some source data and a novel data distribution at test time.
We find that models adapted using TeST significantly improve over baseline test-time adaptation algorithms.
arXiv Detail & Related papers (2022-09-23T07:47:33Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Efficient Test-Time Model Adaptation without Forgetting [60.36499845014649]
Test-time adaptation seeks to tackle potential distribution shifts between training and testing data.
We propose an active sample selection criterion to identify reliable and non-redundant samples.
We also introduce a Fisher regularizer to constrain important model parameters from drastic changes.
arXiv Detail & Related papers (2022-04-06T06:39:40Z) - Continual Test-Time Domain Adaptation [94.51284735268597]
Test-time domain adaptation aims to adapt a source pre-trained model to a target domain without using any source data.
CoTTA is easy to implement and can be readily incorporated in off-the-shelf pre-trained models.
arXiv Detail & Related papers (2022-03-25T11:42:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.