Towards Unified and Effective Domain Generalization
- URL: http://arxiv.org/abs/2310.10008v1
- Date: Mon, 16 Oct 2023 02:05:03 GMT
- Title: Towards Unified and Effective Domain Generalization
- Authors: Yiyuan Zhang, Kaixiong Gong, Xiaohan Ding, Kaipeng Zhang, Fangrui Lv,
Kurt Keutzer, Xiangyu Yue
- Abstract summary: We propose $textbfUniDG$, a novel and $textbfUnified framework for $textbfD$omain $textbfG$eneralization.
Specifically, we encourage models to learn the distribution of test data in an unsupervised manner and impose a penalty regarding the updating step of model parameters.
- Score: 42.579796989908914
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose $\textbf{UniDG}$, a novel and $\textbf{Uni}$fied framework for
$\textbf{D}$omain $\textbf{G}$eneralization that is capable of significantly
enhancing the out-of-distribution generalization performance of foundation
models regardless of their architectures. The core idea of UniDG is to finetune
models during the inference stage, which saves the cost of iterative training.
Specifically, we encourage models to learn the distribution of test data in an
unsupervised manner and impose a penalty regarding the updating step of model
parameters. The penalty term can effectively reduce the catastrophic forgetting
issue as we would like to maximally preserve the valuable knowledge in the
original model. Empirically, across 12 visual backbones, including CNN-, MLP-,
and Transformer-based models, ranging from 1.89M to 303M parameters, UniDG
shows an average accuracy improvement of +5.4% on DomainBed. These performance
results demonstrate the superiority and versatility of UniDG. The code is
publicly available at https://github.com/invictus717/UniDG
Related papers
- Improving Out-of-Distribution Data Handling and Corruption Resistance via Modern Hopfield Networks [0.0]
This study explores the potential of Modern Hopfield Networks (MHN) in improving the ability of computer vision models to handle out-of-distribution data.
We suggest integrating MHN into the baseline models to enhance their robustness.
Our research shows that the proposed integration consistently improves model performance on the MNIST-C dataset.
arXiv Detail & Related papers (2024-08-21T03:26:16Z) - $t^3$-Variational Autoencoder: Learning Heavy-tailed Data with Student's
t and Power Divergence [7.0479532872043755]
$t3$VAE is a modified VAE framework that incorporates Student's t-distributions for the prior, encoder, and decoder.
We show that $t3$VAE significantly outperforms other models on CelebA and imbalanced CIFAR-100 datasets.
arXiv Detail & Related papers (2023-12-02T13:14:28Z) - Generalized Logit Adjustment: Calibrating Fine-tuned Models by Removing Label Bias in Foundation Models [75.9543301303586]
Foundation models like CLIP allow zero-shot transfer on various tasks without additional training data.
Fine-tuning and ensembling are also commonly adopted to better fit the downstream tasks.
However, we argue that prior work has overlooked the inherent biases in foundation models.
arXiv Detail & Related papers (2023-10-12T08:01:11Z) - Universal Domain Adaptation from Foundation Models: A Baseline Study [58.51162198585434]
We make empirical studies of state-of-the-art UniDA methods using foundation models.
We introduce textitCLIP distillation, a parameter-free method specifically designed to distill target knowledge from CLIP models.
Although simple, our method outperforms previous approaches in most benchmark tasks.
arXiv Detail & Related papers (2023-05-18T16:28:29Z) - Better Diffusion Models Further Improve Adversarial Training [97.44991845907708]
It has been recognized that the data generated by the diffusion probabilistic model (DDPM) improves adversarial training.
This paper gives an affirmative answer by employing the most recent diffusion model which has higher efficiency.
Our adversarially trained models achieve state-of-the-art performance on RobustBench using only generated data.
arXiv Detail & Related papers (2023-02-09T13:46:42Z) - Domain Generalization using Pretrained Models without Fine-tuning [25.489714555859944]
Fine-tuning pretrained models is a common practice in domain generalization (DG) tasks.
We propose a novel domain generalization paradigm to better leverage various pretrained models, named specialized ensemble learning for domain generalization (SEDGE)
SEDGE achieves significant performance improvements comparing to strong baselines including state-of-the-art method in DG tasks.
arXiv Detail & Related papers (2022-03-09T09:33:59Z) - Ensemble of Averages: Improving Model Selection and Boosting Performance
in Domain Generalization [63.28279815753543]
In Domain Generalization (DG) settings, models trained on a given set of training domains have notoriously chaotic performance on shifted test domains.
We first show that a simple protocol for averaging model parameters along the optimization path, starting early during training, significantly boosts domain generalizationity.
We show that an ensemble of independently trained models also has a chaotic behavior in the DG setting.
arXiv Detail & Related papers (2021-10-21T00:08:17Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.