Train/Test-Time Adaptation with Retrieval
- URL: http://arxiv.org/abs/2303.14333v1
- Date: Sat, 25 Mar 2023 02:44:57 GMT
- Title: Train/Test-Time Adaptation with Retrieval
- Authors: Luca Zancato, Alessandro Achille, Tian Yu Liu, Matthew Trager,
Pramuditha Perera, Stefano Soatto
- Abstract summary: We introduce Train/Test-Time Adaptation with Retrieval ($rm T3AR$), a method to adapt models both at train and test time.
$rm T3AR$ adapts a given model to the downstream task using refined pseudo-labels and a self-supervised contrastive objective function.
Thanks to the retrieval module, our method gives the user or service provider the possibility to improve model adaptation on the downstream task.
- Score: 129.8579208970529
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce Train/Test-Time Adaptation with Retrieval (${\rm T^3AR}$), a
method to adapt models both at train and test time by means of a retrieval
module and a searchable pool of external samples. Before inference, ${\rm
T^3AR}$ adapts a given model to the downstream task using refined pseudo-labels
and a self-supervised contrastive objective function whose noise distribution
leverages retrieved real samples to improve feature adaptation on the target
data manifold. The retrieval of real images is key to ${\rm T^3AR}$ since it
does not rely solely on synthetic data augmentations to compensate for the lack
of adaptation data, as typically done by other adaptation algorithms.
Furthermore, thanks to the retrieval module, our method gives the user or
service provider the possibility to improve model adaptation on the downstream
task by incorporating further relevant data or to fully remove samples that may
no longer be available due to changes in user preference after deployment.
First, we show that ${\rm T^3AR}$ can be used at training time to improve
downstream fine-grained classification over standard fine-tuning baselines, and
the fewer the adaptation data the higher the relative improvement (up to 13%).
Second, we apply ${\rm T^3AR}$ for test-time adaptation and show that
exploiting a pool of external images at test-time leads to more robust
representations over existing methods on DomainNet-126 and VISDA-C, especially
when few adaptation data are available (up to 8%).
Related papers
- Few-shot Online Anomaly Detection and Segmentation [29.693357653538474]
This paper focuses on addressing the challenging yet practical few-shot online anomaly detection and segmentation (FOADS) task.
Under the FOADS framework, models are trained on a few-shot normal dataset, followed by inspection and improvement of their capabilities by leveraging unlabeled streaming data containing both normal and abnormal samples simultaneously.
In order to achieve improved performance with limited training samples, we employ multi-scale feature embedding extracted from a CNN pre-trained on ImageNet to obtain a robust representation.
arXiv Detail & Related papers (2024-03-27T02:24:00Z) - What, How, and When Should Object Detectors Update in Continually
Changing Test Domains? [34.13756022890991]
Test-time adaptation algorithms have been proposed to adapt the model online while inferring test data.
We propose a novel online adaption approach for object detection in continually changing test domains.
Our approach surpasses baselines on widely used benchmarks, achieving improvements of up to 4.9%p and 7.9%p in mAP.
arXiv Detail & Related papers (2023-12-12T07:13:08Z) - $\Delta$-Patching: A Framework for Rapid Adaptation of Pre-trained
Convolutional Networks without Base Performance Loss [71.46601663956521]
Models pre-trained on large-scale datasets are often fine-tuned to support newer tasks and datasets that arrive over time.
We propose $Delta$-Patching for fine-tuning neural network models in an efficient manner, without the need to store model copies.
Our experiments show that $Delta$-Networks outperform earlier model patching work while only requiring a fraction of parameters to be trained.
arXiv Detail & Related papers (2023-03-26T16:39:44Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - On-the-Fly Test-time Adaptation for Medical Image Segmentation [63.476899335138164]
Adapting the source model to target data distribution at test-time is an efficient solution for the data-shift problem.
We propose a new framework called Adaptive UNet where each convolutional block is equipped with an adaptive batch normalization layer.
During test-time, the model takes in just the new test image and generates a domain code to adapt the features of source model according to the test data.
arXiv Detail & Related papers (2022-03-10T18:51:29Z) - Test-time Batch Statistics Calibration for Covariate Shift [66.7044675981449]
We propose to adapt the deep models to the novel environment during inference.
We present a general formulation $alpha$-BN to calibrate the batch statistics.
We also present a novel loss function to form a unified test time adaptation framework Core.
arXiv Detail & Related papers (2021-10-06T08:45:03Z) - MT3: Meta Test-Time Training for Self-Supervised Test-Time Adaption [69.76837484008033]
An unresolved problem in Deep Learning is the ability of neural networks to cope with domain shifts during test-time.
We combine meta-learning, self-supervision and test-time training to learn to adapt to unseen test distributions.
Our approach significantly improves the state-of-the-art results on the CIFAR-10-Corrupted image classification benchmark.
arXiv Detail & Related papers (2021-03-30T09:33:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.