MixStyle Neural Networks for Domain Generalization and Adaptation
- URL: http://arxiv.org/abs/2107.02053v2
- Date: Fri, 15 Sep 2023 16:42:36 GMT
- Title: MixStyle Neural Networks for Domain Generalization and Adaptation
- Authors: Kaiyang Zhou, Yongxin Yang, Yu Qiao, Tao Xiang
- Abstract summary: MixStyle is a plug-and-play module that can improve domain generalization performance without the need to collect more data or increase model capacity.
Our experiments show that MixStyle can significantly boost out-of-distribution generalization performance across a wide range of tasks including image recognition, instance retrieval and reinforcement learning.
- Score: 122.36901703868321
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural networks do not generalize well to unseen data with domain shifts -- a
longstanding problem in machine learning and AI. To overcome the problem, we
propose MixStyle, a simple plug-and-play, parameter-free module that can
improve domain generalization performance without the need to collect more data
or increase model capacity. The design of MixStyle is simple: it mixes the
feature statistics of two random instances in a single forward pass during
training. The idea is grounded by the finding from recent style transfer
research that feature statistics capture image style information, which
essentially defines visual domains. Therefore, mixing feature statistics can be
seen as an efficient way to synthesize new domains in the feature space, thus
achieving data augmentation. MixStyle is easy to implement with a few lines of
code, does not require modification to training objectives, and can fit a
variety of learning paradigms including supervised domain generalization,
semi-supervised domain generalization, and unsupervised domain adaptation. Our
experiments show that MixStyle can significantly boost out-of-distribution
generalization performance across a wide range of tasks including image
recognition, instance retrieval and reinforcement learning.
Related papers
- StyDeSty: Min-Max Stylization and Destylization for Single Domain Generalization [85.18995948334592]
Single domain generalization (single DG) aims at learning a robust model generalizable to unseen domains from only one training domain.
State-of-the-art approaches have mostly relied on data augmentations, such as adversarial perturbation and style enhancement, to synthesize new data.
We propose emphStyDeSty, which explicitly accounts for the alignment of the source and pseudo domains in the process of data augmentation.
arXiv Detail & Related papers (2024-06-01T02:41:34Z) - TFS-ViT: Token-Level Feature Stylization for Domain Generalization [17.82872117103924]
Vision Transformers (ViTs) have shown outstanding performance for a broad range of computer vision tasks.
This paper presents a first Token-level Feature Stylization (TFS-ViT) approach for domain generalization.
Our approach transforms token features by mixing the normalization statistics of images from different domains.
arXiv Detail & Related papers (2023-03-28T03:00:28Z) - Adversarial Style Augmentation for Domain Generalization [41.72506801753435]
We introduce a novel Adrial Style Augmentation (ASA) method, which explores broader style spaces by generating more effective statistics perturbation.
To facilitate the application of ASA, we design a simple yet effective module, namely AdvStyle, which instantiates the ASA method in a plug-and-play manner.
Our method significantly outperforms its competitors on the PACS dataset under the single source generalization setting.
arXiv Detail & Related papers (2023-01-30T03:52:16Z) - Domain-incremental Cardiac Image Segmentation with Style-oriented Replay
and Domain-sensitive Feature Whitening [67.6394526631557]
M&Ms should incrementally learn from each incoming dataset and progressively update with improved functionality as time goes by.
In medical scenarios, this is particularly challenging as accessing or storing past data is commonly not allowed due to data privacy.
We propose a novel domain-incremental learning framework to recover past domain inputs first and then regularly replay them during model optimization.
arXiv Detail & Related papers (2022-11-09T13:07:36Z) - FIXED: Frustratingly Easy Domain Generalization with Mixup [53.782029033068675]
Domain generalization (DG) aims to learn a generalizable model from multiple training domains such that it can perform well on unseen target domains.
A popular strategy is to augment training data to benefit generalization through methods such as Mixupcitezhang 2018mixup.
We propose a simple yet effective enhancement for Mixup-based DG, namely domain-invariant Feature mIXup (FIX)
Our approach significantly outperforms nine state-of-the-art related methods, beating the best performing baseline by 6.5% on average in terms of test accuracy.
arXiv Detail & Related papers (2022-11-07T09:38:34Z) - Semi-supervised Meta-learning with Disentanglement for
Domain-generalised Medical Image Segmentation [15.351113774542839]
Generalising models to new data from new centres (termed here domains) remains a challenge.
We propose a novel semi-supervised meta-learning framework with disentanglement.
We show that the proposed method is robust on different segmentation tasks and achieves state-of-the-art generalisation performance on two public benchmarks.
arXiv Detail & Related papers (2021-06-24T19:50:07Z) - Domain Generalization with MixStyle [120.52367818581608]
Domain generalization aims to address this problem by learning from a set of source domains a model that is generalizable to any unseen domain.
Our method, termed MixStyle, is motivated by the observation that visual domain is closely related to image style.
MixStyle fits into mini-batch training perfectly and is extremely easy to implement.
arXiv Detail & Related papers (2021-04-05T16:58:09Z) - i-Mix: A Domain-Agnostic Strategy for Contrastive Representation
Learning [117.63815437385321]
We propose i-Mix, a simple yet effective domain-agnostic regularization strategy for improving contrastive representation learning.
In experiments, we demonstrate that i-Mix consistently improves the quality of learned representations across domains.
arXiv Detail & Related papers (2020-10-17T23:32:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.