TTN: A Domain-Shift Aware Batch Normalization in Test-Time Adaptation
- URL: http://arxiv.org/abs/2302.05155v1
- Date: Fri, 10 Feb 2023 10:25:29 GMT
- Title: TTN: A Domain-Shift Aware Batch Normalization in Test-Time Adaptation
- Authors: Hyesu Lim, Byeonggeun Kim, Jaegul Choo, Sungha Choi
- Abstract summary: Recent test-time adaptation methods heavily rely on transductive batch normalization (TBN)
Adopting TBN that employs test batch statistics mitigates the performance degradation caused by the domain shift.
We present a new test-time normalization (TTN) method that interpolates the statistics by adjusting the importance between CBN and TBN according to the domain-shift sensitivity of each BN layer.
- Score: 28.63285970880039
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper proposes a novel batch normalization strategy for test-time
adaptation. Recent test-time adaptation methods heavily rely on the modified
batch normalization, i.e., transductive batch normalization (TBN), which
calculates the mean and the variance from the current test batch rather than
using the running mean and variance obtained from the source data, i.e.,
conventional batch normalization (CBN). Adopting TBN that employs test batch
statistics mitigates the performance degradation caused by the domain shift.
However, re-estimating normalization statistics using test data depends on
impractical assumptions that a test batch should be large enough and be drawn
from i.i.d. stream, and we observed that the previous methods with TBN show
critical performance drop without the assumptions. In this paper, we identify
that CBN and TBN are in a trade-off relationship and present a new test-time
normalization (TTN) method that interpolates the statistics by adjusting the
importance between CBN and TBN according to the domain-shift sensitivity of
each BN layer. Our proposed TTN improves model robustness to shifted domains
across a wide range of batch sizes and in various realistic evaluation
scenarios. TTN is widely applicable to other test-time adaptation methods that
rely on updating model parameters via backpropagation. We demonstrate that
adopting TTN further improves their performance and achieves state-of-the-art
performance in various standard benchmarks.
Related papers
- Optimization-Free Test-Time Adaptation for Cross-Person Activity
Recognition [30.350005654271868]
Test-Time Adaptation aims to utilize the test stream to adjust predictions in real-time inference.
High computational cost makes it intractable to run on resource-constrained edge devices.
We propose an Optimization-Free Test-Time Adaptation framework for sensor-based HAR.
arXiv Detail & Related papers (2023-10-28T02:20:33Z) - On Pitfalls of Test-Time Adaptation [82.8392232222119]
Test-Time Adaptation (TTA) has emerged as a promising approach for tackling the robustness challenge under distribution shifts.
We present TTAB, a test-time adaptation benchmark that encompasses ten state-of-the-art algorithms, a diverse array of distribution shifts, and two evaluation protocols.
arXiv Detail & Related papers (2023-06-06T09:35:29Z) - Robust Continual Test-time Adaptation: Instance-aware BN and
Prediction-balanced Memory [58.72445309519892]
We present a new test-time adaptation scheme that is robust against non-i.i.d. test data streams.
Our novelty is mainly two-fold: (a) Instance-Aware Batch Normalization (IABN) that corrects normalization for out-of-distribution samples, and (b) Prediction-balanced Reservoir Sampling (PBRS) that simulates i.i.d. data stream from non-i.i.d. stream in a class-balanced manner.
arXiv Detail & Related papers (2022-08-10T03:05:46Z) - Test-time Batch Normalization [61.292862024903584]
Deep neural networks often suffer the data distribution shift between training and testing.
We revisit the batch normalization (BN) in the training process and reveal two key insights benefiting test-time optimization.
We propose a novel test-time BN layer design, GpreBN, which is optimized during testing by minimizing Entropy loss.
arXiv Detail & Related papers (2022-05-20T14:33:39Z) - Test-time Batch Statistics Calibration for Covariate Shift [66.7044675981449]
We propose to adapt the deep models to the novel environment during inference.
We present a general formulation $alpha$-BN to calibrate the batch statistics.
We also present a novel loss function to form a unified test time adaptation framework Core.
arXiv Detail & Related papers (2021-10-06T08:45:03Z) - Double Forward Propagation for Memorized Batch Normalization [68.34268180871416]
Batch Normalization (BN) has been a standard component in designing deep neural networks (DNNs)
We propose a memorized batch normalization (MBN) which considers multiple recent batches to obtain more accurate and robust statistics.
Compared to related methods, the proposed MBN exhibits consistent behaviors in both training and inference.
arXiv Detail & Related papers (2020-10-10T08:48:41Z) - PowerNorm: Rethinking Batch Normalization in Transformers [96.14956636022957]
normalization method for neural network (NN) models used in Natural Language Processing (NLP) is layer normalization (LN)
LN is preferred due to the empirical observation that a (naive/vanilla) use of BN leads to significant performance degradation for NLP tasks.
We propose Power Normalization (PN), a novel normalization scheme that resolves this issue.
arXiv Detail & Related papers (2020-03-17T17:50:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.