Learning to Augment via Implicit Differentiation for Domain
Generalization
- URL: http://arxiv.org/abs/2210.14271v1
- Date: Tue, 25 Oct 2022 18:51:51 GMT
- Title: Learning to Augment via Implicit Differentiation for Domain
Generalization
- Authors: Tingwei Wang, Da Li, Kaiyang Zhou, Tao Xiang and Yi-Zhe Song
- Abstract summary: Domain generalization (DG) aims to overcome the problem by leveraging multiple source domains to learn a domain-generalizable model.
In this paper, we propose a novel augmentation-based DG approach, dubbed AugLearn.
AugLearn shows effectiveness on three standard DG benchmarks, PACS, Office-Home and Digits-DG.
- Score: 107.9666735637355
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine learning models are intrinsically vulnerable to domain shift between
training and testing data, resulting in poor performance in novel domains.
Domain generalization (DG) aims to overcome the problem by leveraging multiple
source domains to learn a domain-generalizable model. In this paper, we propose
a novel augmentation-based DG approach, dubbed AugLearn. Different from
existing data augmentation methods, our AugLearn views a data augmentation
module as hyper-parameters of a classification model and optimizes the module
together with the model via meta-learning. Specifically, at each training step,
AugLearn (i) divides source domains into a pseudo source and a pseudo target
set, and (ii) trains the augmentation module in such a way that the augmented
(synthetic) images can make the model generalize well on the pseudo target set.
Moreover, to overcome the expensive second-order gradient computation during
meta-learning, we formulate an efficient joint training algorithm, for both the
augmentation module and the classification model, based on the implicit
function theorem. With the flexibility of augmenting data in both time and
frequency spaces, AugLearn shows effectiveness on three standard DG benchmarks,
PACS, Office-Home and Digits-DG.
Related papers
- StyDeSty: Min-Max Stylization and Destylization for Single Domain Generalization [85.18995948334592]
Single domain generalization (single DG) aims at learning a robust model generalizable to unseen domains from only one training domain.
State-of-the-art approaches have mostly relied on data augmentations, such as adversarial perturbation and style enhancement, to synthesize new data.
We propose emphStyDeSty, which explicitly accounts for the alignment of the source and pseudo domains in the process of data augmentation.
arXiv Detail & Related papers (2024-06-01T02:41:34Z) - SETA: Semantic-Aware Token Augmentation for Domain Generalization [27.301312891532277]
Domain generalization (DG) aims to enhance the model against domain shifts without accessing target domains.
Prior CNN-based augmentation methods on token-based models are suboptimal due to the lack of incentivizing the model to learn holistic shape information.
We propose the SEmantic-aware Token Augmentation (SETA) method, which transforms features by perturbing local edge cues while preserving global shape features.
arXiv Detail & Related papers (2024-03-18T13:50:35Z) - CNN Feature Map Augmentation for Single-Source Domain Generalization [6.053629733936548]
Domain Generalization (DG) has gained significant traction during the past few years.
The goal in DG is to produce models which continue to perform well when presented with data distributions different from the ones available during training.
We propose an alternative regularization technique for convolutional neural network architectures in the single-source DG image classification setting.
arXiv Detail & Related papers (2023-05-26T08:48:17Z) - Dynamic Domain Generalization [10.252262302724608]
We develop a brand-new DG variant, namely Dynamic Domain Generalization (DDG), in which the model learns to twist the network parameters to adapt the data from different domains.
Specifically, we leverage a meta-adjuster to twist the network parameters based on the static model with respect to different data from different domains.
In this way, the static model is optimized to learn domain-shared features, while the meta-adjuster is designed to learn domain-specific features.
arXiv Detail & Related papers (2022-05-27T11:29:03Z) - Source-Free Open Compound Domain Adaptation in Semantic Segmentation [99.82890571842603]
In SF-OCDA, only the source pre-trained model and the target data are available to learn the target model.
We propose the Cross-Patch Style Swap (CPSS) to diversify samples with various patch styles in the feature-level.
Our method produces state-of-the-art results on the C-Driving dataset.
arXiv Detail & Related papers (2021-06-07T08:38:41Z) - Learning to Generalize Unseen Domains via Memory-based Multi-Source
Meta-Learning for Person Re-Identification [59.326456778057384]
We propose the Memory-based Multi-Source Meta-Learning framework to train a generalizable model for unseen domains.
We also present a meta batch normalization layer (MetaBN) to diversify meta-test features.
Experiments demonstrate that our M$3$L can effectively enhance the generalization ability of the model for unseen domains.
arXiv Detail & Related papers (2020-12-01T11:38:16Z) - Discriminative Adversarial Domain Generalization with Meta-learning
based Cross-domain Validation [9.265557367859637]
Domain Generalization (DG) techniques aim to enhance such generalization capability of machine learning models.
We propose Discriminative Adversarial Domain Generalization (DADG) with meta-learning-based cross-domain validation.
Results show DADG consistently outperforms a strong baseline DeepAll, and outperforms the other existing DG algorithms in most of the evaluation cases.
arXiv Detail & Related papers (2020-11-01T07:48:16Z) - Adversarial Bipartite Graph Learning for Video Domain Adaptation [50.68420708387015]
Domain adaptation techniques, which focus on adapting models between distributionally different domains, are rarely explored in the video recognition area.
Recent works on visual domain adaptation which leverage adversarial learning to unify the source and target video representations are not highly effective on the videos.
This paper proposes an Adversarial Bipartite Graph (ABG) learning framework which directly models the source-target interactions.
arXiv Detail & Related papers (2020-07-31T03:48:41Z) - Learning to Generate Novel Domains for Domain Generalization [115.21519842245752]
This paper focuses on the task of learning from multiple source domains a model that generalizes well to unseen domains.
We employ a data generator to synthesize data from pseudo-novel domains to augment the source domains.
Our method, L2A-OT, outperforms current state-of-the-art DG methods on four benchmark datasets.
arXiv Detail & Related papers (2020-07-07T09:34:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.