Towards Open-Set Test-Time Adaptation Utilizing the Wisdom of Crowds in
Entropy Minimization
- URL: http://arxiv.org/abs/2308.06879v2
- Date: Mon, 4 Sep 2023 08:37:03 GMT
- Title: Towards Open-Set Test-Time Adaptation Utilizing the Wisdom of Crowds in
Entropy Minimization
- Authors: Jungsoo Lee, Debasmit Das, Jaegul Choo, Sungha Choi
- Abstract summary: Test-time adaptation (TTA) methods rely on the model's predictions to adapt the source pretrained model to the unlabeled target domain.
We propose a simple yet effective sample selection method inspired by the following crucial empirical finding.
- Score: 47.61333493671805
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Test-time adaptation (TTA) methods, which generally rely on the model's
predictions (e.g., entropy minimization) to adapt the source pretrained model
to the unlabeled target domain, suffer from noisy signals originating from 1)
incorrect or 2) open-set predictions. Long-term stable adaptation is hampered
by such noisy signals, so training models without such error accumulation is
crucial for practical TTA. To address these issues, including open-set TTA, we
propose a simple yet effective sample selection method inspired by the
following crucial empirical finding. While entropy minimization compels the
model to increase the probability of its predicted label (i.e., confidence
values), we found that noisy samples rather show decreased confidence values.
To be more specific, entropy minimization attempts to raise the confidence
values of an individual sample's prediction, but individual confidence values
may rise or fall due to the influence of signals from numerous other
predictions (i.e., wisdom of crowds). Due to this fact, noisy signals
misaligned with such 'wisdom of crowds', generally found in the correct
signals, fail to raise the individual confidence values of wrong samples,
despite attempts to increase them. Based on such findings, we filter out the
samples whose confidence values are lower in the adapted model than in the
original model, as they are likely to be noisy. Our method is widely applicable
to existing TTA methods and improves their long-term adaptation performance in
both image classification (e.g., 49.4% reduced error rates with TENT) and
semantic segmentation (e.g., 11.7% gain in mIoU with TENT).
Related papers
- COME: Test-time adaption by Conservatively Minimizing Entropy [45.689829178140634]
Conservatively Minimize the Entropy (COME) is a drop-in replacement of traditional entropy (EM)
COME explicitly models the uncertainty by characterizing a Dirichlet prior distribution over model predictions.
We show that COME achieves state-of-the-art performance on commonly used benchmarks.
arXiv Detail & Related papers (2024-10-12T09:20:06Z) - Uncertainty-Calibrated Test-Time Model Adaptation without Forgetting [55.17761802332469]
Test-time adaptation (TTA) seeks to tackle potential distribution shifts between training and test data by adapting a given model w.r.t. any test sample.
Prior methods perform backpropagation for each test sample, resulting in unbearable optimization costs to many applications.
We propose an Efficient Anti-Forgetting Test-Time Adaptation (EATA) method which develops an active sample selection criterion to identify reliable and non-redundant samples.
arXiv Detail & Related papers (2024-03-18T05:49:45Z) - Decoupled Prototype Learning for Reliable Test-Time Adaptation [50.779896759106784]
Test-time adaptation (TTA) is a task that continually adapts a pre-trained source model to the target domain during inference.
One popular approach involves fine-tuning model with cross-entropy loss according to estimated pseudo-labels.
This study reveals that minimizing the classification error of each sample causes the cross-entropy loss's vulnerability to label noise.
We propose a novel Decoupled Prototype Learning (DPL) method that features prototype-centric loss computation.
arXiv Detail & Related papers (2024-01-15T03:33:39Z) - REALM: Robust Entropy Adaptive Loss Minimization for Improved
Single-Sample Test-Time Adaptation [5.749155230209001]
Fully-test-time adaptation (F-TTA) can mitigate performance loss due to distribution shifts between train and test data.
We present a general framework for improving robustness of F-TTA to noisy samples, inspired by self-paced learning and robust loss functions.
arXiv Detail & Related papers (2023-09-07T18:44:58Z) - Calibrating Deep Neural Networks using Explicit Regularisation and
Dynamic Data Pruning [25.982037837953268]
Deep neural networks (DNN) are prone to miscalibrated predictions, often exhibiting a mismatch between the predicted output and the associated confidence scores.
We propose a novel regularization technique that can be used with classification losses, leading to state-of-the-art calibrated predictions at test time.
arXiv Detail & Related papers (2022-12-20T05:34:58Z) - Efficient Test-Time Model Adaptation without Forgetting [60.36499845014649]
Test-time adaptation seeks to tackle potential distribution shifts between training and testing data.
We propose an active sample selection criterion to identify reliable and non-redundant samples.
We also introduce a Fisher regularizer to constrain important model parameters from drastic changes.
arXiv Detail & Related papers (2022-04-06T06:39:40Z) - MEMO: Test Time Robustness via Adaptation and Augmentation [131.28104376280197]
We study the problem of test time robustification, i.e., using the test input to improve model robustness.
Recent prior works have proposed methods for test time adaptation, however, they each introduce additional assumptions.
We propose a simple approach that can be used in any test setting where the model is probabilistic and adaptable.
arXiv Detail & Related papers (2021-10-18T17:55:11Z) - Confidence Adaptive Regularization for Deep Learning with Noisy Labels [2.0349696181833337]
Recent studies on the memorization effects of deep neural networks on noisy labels show that the networks first fit the correctly-labeled training samples before memorizing the mislabeled samples.
Motivated by this early-learning phenomenon, we propose a novel method to prevent memorization of the mislabeled samples.
We provide the theoretical analysis and conduct the experiments on synthetic and real-world datasets, demonstrating that our approach achieves comparable results to the state-of-the-art methods.
arXiv Detail & Related papers (2021-08-18T15:51:25Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.