Beyond Entropy: Style Transfer Guided Single Image Continual Test-Time
Adaptation
- URL: http://arxiv.org/abs/2311.18270v1
- Date: Thu, 30 Nov 2023 06:14:24 GMT
- Title: Beyond Entropy: Style Transfer Guided Single Image Continual Test-Time
Adaptation
- Authors: Younggeol Cho, Youngrae Kim, Dongman Lee
- Abstract summary: We present BESTTA, a novel single image continual test-time adaptation method guided by style transfer.
We demonstrate that BESTTA effectively adapts to the continually changing target environment, leveraging only a single image.
Remarkably, despite training only two parameters in a BeIN layer consuming the least memory, BESTTA outperforms existing state-of-the-art methods in terms of performance.
- Score: 1.6497679785422956
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continual test-time adaptation (cTTA) methods are designed to facilitate the
continual adaptation of models to dynamically changing real-world environments
where computational resources are limited. Due to this inherent limitation,
existing approaches fail to simultaneously achieve accuracy and efficiency. In
detail, when using a single image, the instability caused by batch
normalization layers and entropy loss significantly destabilizes many existing
methods in real-world cTTA scenarios. To overcome these challenges, we present
BESTTA, a novel single image continual test-time adaptation method guided by
style transfer, which enables stable and efficient adaptation to the target
environment by transferring the style of the input image to the source style.
To implement the proposed method, we devise BeIN, a simple yet powerful
normalization method, along with the style-guided losses. We demonstrate that
BESTTA effectively adapts to the continually changing target environment,
leveraging only a single image on both semantic segmentation and image
classification tasks. Remarkably, despite training only two parameters in a
BeIN layer consuming the least memory, BESTTA outperforms existing
state-of-the-art methods in terms of performance.
Related papers
- PostEdit: Posterior Sampling for Efficient Zero-Shot Image Editing [63.38854614997581]
We introduce PostEdit, a method that incorporates a posterior scheme to govern the diffusion sampling process.
The proposed PostEdit achieves state-of-the-art editing performance while accurately preserving unedited regions.
The method is both inversion- and training-free, necessitating approximately 1.5 seconds and 18 GB of GPU memory to generate high-quality results.
arXiv Detail & Related papers (2024-10-07T09:04:50Z) - OSSA: Unsupervised One-Shot Style Adaptation [41.71187047855695]
We introduce One-Shot Style Adaptation (OSSA), a novel unsupervised domain adaptation method for object detection.
OSSA generates diverse target styles by perturbing the style statistics derived from a single target image.
We show that OSSA establishes a new state-of-the-art among one-shot domain adaptation methods by a significant margin.
arXiv Detail & Related papers (2024-10-01T17:43:57Z) - Towards Continual Learning Desiderata via HSIC-Bottleneck
Orthogonalization and Equiangular Embedding [55.107555305760954]
We propose a conceptually simple yet effective method that attributes forgetting to layer-wise parameter overwriting and the resulting decision boundary distortion.
Our method achieves competitive accuracy performance, even with absolute superiority of zero exemplar buffer and 1.02x the base model.
arXiv Detail & Related papers (2024-01-17T09:01:29Z) - Test-Time Training for Semantic Segmentation with Output Contrastive
Loss [12.535720010867538]
Deep learning-based segmentation models have achieved impressive performance on public benchmarks, but generalizing well to unseen environments remains a major challenge.
This paper introduces Contrastive Loss (OCL), known for its capability to learn robust and generalized representations, to stabilize the adaptation process.
Our method excels even when applied to models initially pre-trained using domain adaptation methods on test domain data, showcasing its resilience and adaptability.
arXiv Detail & Related papers (2023-11-14T03:13:47Z) - Effective Restoration of Source Knowledge in Continual Test Time
Adaptation [44.17577480511772]
This paper introduces an unsupervised domain change detection method that is capable of identifying domain shifts in dynamic environments.
By restoring the knowledge from the source, it effectively corrects the negative consequences arising from the gradual deterioration of model parameters.
We perform extensive experiments on benchmark datasets to demonstrate the superior performance of our method compared to state-of-the-art adaptation methods.
arXiv Detail & Related papers (2023-11-08T19:21:48Z) - AR-TTA: A Simple Method for Real-World Continual Test-Time Adaptation [1.4530711901349282]
We propose to validate test-time adaptation methods using datasets for autonomous driving, namely CLAD-C and SHIFT.
We observe that current test-time adaptation methods struggle to effectively handle varying degrees of domain shift.
We enhance the well-established self-training framework by incorporating a small memory buffer to increase model stability.
arXiv Detail & Related papers (2023-09-18T19:34:23Z) - REALM: Robust Entropy Adaptive Loss Minimization for Improved
Single-Sample Test-Time Adaptation [5.749155230209001]
Fully-test-time adaptation (F-TTA) can mitigate performance loss due to distribution shifts between train and test data.
We present a general framework for improving robustness of F-TTA to noisy samples, inspired by self-paced learning and robust loss functions.
arXiv Detail & Related papers (2023-09-07T18:44:58Z) - Improving Diversity in Zero-Shot GAN Adaptation with Semantic Variations [61.132408427908175]
zero-shot GAN adaptation aims to reuse well-trained generators to synthesize images of an unseen target domain.
With only a single representative text feature instead of real images, the synthesized images gradually lose diversity.
We propose a novel method to find semantic variations of the target text in the CLIP space.
arXiv Detail & Related papers (2023-08-21T08:12:28Z) - Condition-Invariant Semantic Segmentation [77.10045325743644]
We implement Condition-Invariant Semantic (CISS) on the current state-of-the-art domain adaptation architecture.
Our method achieves the second-best performance on the normal-to-adverse Cityscapes$to$ACDC benchmark.
CISS is shown to generalize well to domains unseen during training, such as BDD100K-night and ACDC-night.
arXiv Detail & Related papers (2023-05-27T03:05:07Z) - CCPL: Contrastive Coherence Preserving Loss for Versatile Style Transfer [58.020470877242865]
We devise a universally versatile style transfer method capable of performing artistic, photo-realistic, and video style transfer jointly.
We make a mild and reasonable assumption that global inconsistency is dominated by local inconsistencies and devise a generic Contrastive Coherence Preserving Loss (CCPL) applied to local patches.
CCPL can preserve the coherence of the content source during style transfer without degrading stylization.
arXiv Detail & Related papers (2022-07-11T12:09:41Z) - Towards Full-to-Empty Room Generation with Structure-Aware Feature
Encoding and Soft Semantic Region-Adaptive Normalization [67.64622529651677]
We propose a simple yet effective adjusted fully differentiable soft semantic region-adaptive normalization module (softSEAN) block.
Our approach besides the advantages of mitigating training complexity and non-differentiability issues surpasses the compared methods both quantitatively and qualitatively.
Our softSEAN block can be used as a drop-in module for existing discriminative and generative models.
arXiv Detail & Related papers (2021-12-10T09:00:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.