Learning to Learn Domain-invariant Parameters for Domain Generalization
- URL: http://arxiv.org/abs/2211.04582v1
- Date: Fri, 4 Nov 2022 07:19:34 GMT
- Title: Learning to Learn Domain-invariant Parameters for Domain Generalization
- Authors: Feng Hou, Yao Zhang, Yang Liu, Jin Yuan, Cheng Zhong, Yang Zhang,
Zhongchao Shi, Jianping Fan, Zhiqiang He
- Abstract summary: Domain generalization (DG) aims to overcome this issue by capturing domain-invariant representations from source domains.
We propose two modules of Domain Decoupling and Combination (DDC) and Domain-invariance-guided Backpropagation (DIGB)
Our proposed method has achieved state-of-the-art performance with strong generalization capability.
- Score: 29.821634033299855
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to domain shift, deep neural networks (DNNs) usually fail to generalize
well on unknown test data in practice. Domain generalization (DG) aims to
overcome this issue by capturing domain-invariant representations from source
domains. Motivated by the insight that only partial parameters of DNNs are
optimized to extract domain-invariant representations, we expect a general
model that is capable of well perceiving and emphatically updating such
domain-invariant parameters. In this paper, we propose two modules of Domain
Decoupling and Combination (DDC) and Domain-invariance-guided Backpropagation
(DIGB), which can encourage such general model to focus on the parameters that
have a unified optimization direction between pairs of contrastive samples. Our
extensive experiments on two benchmarks have demonstrated that our proposed
method has achieved state-of-the-art performance with strong generalization
capability.
Related papers
- Bayesian Domain Invariant Learning via Posterior Generalization of
Parameter Distributions [29.018103152856792]
PosTerior Generalization (PTG) shows competitive performance on various domain generalization benchmarks on DomainBed.
PTG fully exploits variational inference to approximate parameter distributions, including the invariant posterior and the posteriors on training domains.
arXiv Detail & Related papers (2023-10-25T01:17:08Z) - Adaptive Domain Generalization via Online Disagreement Minimization [17.215683606365445]
Domain Generalization aims to safely transfer a model to unseen target domains.
AdaODM adaptively modifies the source model at test time for different target domains.
Results show AdaODM stably improves the generalization capacity on unseen domains.
arXiv Detail & Related papers (2022-08-03T11:51:11Z) - Compound Domain Generalization via Meta-Knowledge Encoding [55.22920476224671]
We introduce Style-induced Domain-specific Normalization (SDNorm) to re-normalize the multi-modal underlying distributions.
We harness the prototype representations, the centroids of classes, to perform relational modeling in the embedding space.
Experiments on four standard Domain Generalization benchmarks reveal that COMEN exceeds the state-of-the-art performance without the need of domain supervision.
arXiv Detail & Related papers (2022-03-24T11:54:59Z) - Domain Generalisation for Object Detection under Covariate and Concept Shift [10.32461766065764]
Domain generalisation aims to promote the learning of domain-invariant features while suppressing domain-specific features.
An approach to domain generalisation for object detection is proposed, the first such approach applicable to any object detection architecture.
arXiv Detail & Related papers (2022-03-10T11:14:18Z) - TAL: Two-stream Adaptive Learning for Generalizable Person
Re-identification [115.31432027711202]
We argue that both domain-specific and domain-invariant features are crucial for improving the generalization ability of re-id models.
We name two-stream adaptive learning (TAL) to simultaneously model these two kinds of information.
Our framework can be applied to both single-source and multi-source domain generalization tasks.
arXiv Detail & Related papers (2021-11-29T01:27:42Z) - Invariant Information Bottleneck for Domain Generalization [39.62337297660974]
We propose a novel algorithm that learns a minimally sufficient representation that is invariant across training and testing domains.
By minimizing the mutual information between the representation and inputs, IIB alleviates its reliance on pseudo-invariant features.
The results show that IIB outperforms invariant learning baseline (e.g. IRM) by an average of 2.8% and 3.8% accuracy over two evaluation metrics.
arXiv Detail & Related papers (2021-06-11T12:12:40Z) - Model-Based Domain Generalization [96.84818110323518]
We propose a novel approach for the domain generalization problem called Model-Based Domain Generalization.
Our algorithms beat the current state-of-the-art methods on the very-recently-proposed WILDS benchmark by up to 20 percentage points.
arXiv Detail & Related papers (2021-02-23T00:59:02Z) - Dual Distribution Alignment Network for Generalizable Person
Re-Identification [174.36157174951603]
Domain generalization (DG) serves as a promising solution to handle person Re-Identification (Re-ID)
We present a Dual Distribution Alignment Network (DDAN) which handles this challenge by selectively aligning distributions of multiple source domains.
We evaluate our DDAN on a large-scale Domain Generalization Re-ID (DG Re-ID) benchmark.
arXiv Detail & Related papers (2020-07-27T00:08:07Z) - Learning to Learn with Variational Information Bottleneck for Domain
Generalization [128.90691697063616]
Domain generalization models learn to generalize to previously unseen domains, but suffer from prediction uncertainty and domain shift.
We introduce a probabilistic meta-learning model for domain generalization, in which parameters shared across domains are modeled as distributions.
To deal with domain shift, we learn domain-invariant representations by the proposed principle of meta variational information bottleneck, we call MetaVIB.
arXiv Detail & Related papers (2020-07-15T12:05:52Z) - Bi-Directional Generation for Unsupervised Domain Adaptation [61.73001005378002]
Unsupervised domain adaptation facilitates the unlabeled target domain relying on well-established source domain information.
Conventional methods forcefully reducing the domain discrepancy in the latent space will result in the destruction of intrinsic data structure.
We propose a Bi-Directional Generation domain adaptation model with consistent classifiers interpolating two intermediate domains to bridge source and target domains.
arXiv Detail & Related papers (2020-02-12T09:45:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.