Unknown Domain Inconsistency Minimization for Domain Generalization
- URL: http://arxiv.org/abs/2403.07329v1
- Date: Tue, 12 Mar 2024 05:29:48 GMT
- Title: Unknown Domain Inconsistency Minimization for Domain Generalization
- Authors: Seungjae Shin, HeeSun Bae, Byeonghu Na, Yoon-Yeong Kim and Il-Chul
Moon
- Abstract summary: This paper introduces an objective rooted in both parameter and data perturbed regions for domain generalization, coined Unknown Domain Inconsistency Minimization (UDIM)
UDIM reduces the loss landscape inconsistency between source domain and unknown domains.
In an empirical aspect, UDIM consistently outperforms SAM variants across multiple DG benchmark datasets.
- Score: 18.58931160403153
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The objective of domain generalization (DG) is to enhance the transferability
of the model learned from a source domain to unobserved domains. To prevent
overfitting to a specific domain, Sharpness-Aware Minimization (SAM) reduces
source domain's loss sharpness. Although SAM variants have delivered
significant improvements in DG, we highlight that there's still potential for
improvement in generalizing to unknown domains through the exploration on data
space. This paper introduces an objective rooted in both parameter and data
perturbed regions for domain generalization, coined Unknown Domain
Inconsistency Minimization (UDIM). UDIM reduces the loss landscape
inconsistency between source domain and unknown domains. As unknown domains are
inaccessible, these domains are empirically crafted by perturbing instances
from the source domain dataset. In particular, by aligning the loss landscape
acquired in the source domain to the loss landscape of perturbed domains, we
expect to achieve generalization grounded on these flat minima for the unknown
domains. Theoretically, we validate that merging SAM optimization with the UDIM
objective establishes an upper bound for the true objective of the DG task. In
an empirical aspect, UDIM consistently outperforms SAM variants across multiple
DG benchmark datasets. Notably, UDIM shows statistically significant
improvements in scenarios with more restrictive domain information,
underscoring UDIM's generalization capability in unseen domains. Our code is
available at \url{https://github.com/SJShin-AI/UDIM}.
Related papers
- Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization [71.40801206714382]
We propose an attention-refocusing scheme, called Simulate, Refocus and Ensemble (SRE)<n>SRE learns to reduce the domain shift by aligning the attention maps in CLIP via attention refocusing.<n>Experiments on several datasets demonstrate that SRE generally achieves better results than state-of-the-art methods.
arXiv Detail & Related papers (2025-07-17T07:20:32Z) - Disentangling Masked Autoencoders for Unsupervised Domain Generalization [57.56744870106124]
Unsupervised domain generalization is fast gaining attention but is still far from well-studied.
Disentangled Masked Auto (DisMAE) aims to discover the disentangled representations that faithfully reveal intrinsic features.
DisMAE co-trains the asymmetric dual-branch architecture with semantic and lightweight variation encoders.
arXiv Detail & Related papers (2024-07-10T11:11:36Z) - Overcoming Data Inequality across Domains with Semi-Supervised Domain
Generalization [4.921899151930171]
We propose a novel algorithm, ProUD, which can effectively learn domain-invariant features via domain-aware prototypes.
Our experiments on three different benchmark datasets demonstrate the effectiveness of ProUD.
arXiv Detail & Related papers (2024-03-08T10:49:37Z) - Domain Generalization via Causal Adjustment for Cross-Domain Sentiment
Analysis [59.73582306457387]
We focus on the problem of domain generalization for cross-domain sentiment analysis.
We propose a backdoor adjustment-based causal model to disentangle the domain-specific and domain-invariant representations.
A series of experiments show the great performance and robustness of our model.
arXiv Detail & Related papers (2024-02-22T13:26:56Z) - Make the U in UDA Matter: Invariant Consistency Learning for
Unsupervised Domain Adaptation [86.61336696914447]
We dub our approach "Invariant CONsistency learning" (ICON)
We propose to make the U in Unsupervised DA matter by giving equal status to the two domains.
ICON achieves the state-of-the-art performance on the classic UDA benchmarks: Office-Home and VisDA-2017, and outperforms all the conventional methods on the challenging WILDS 2.0 benchmark.
arXiv Detail & Related papers (2023-09-22T09:43:32Z) - Domain-Agnostic Prior for Transfer Semantic Segmentation [197.9378107222422]
Unsupervised domain adaptation (UDA) is an important topic in the computer vision community.
We present a mechanism that regularizes cross-domain representation learning with a domain-agnostic prior (DAP)
Our research reveals that UDA benefits much from better proxies, possibly from other data modalities.
arXiv Detail & Related papers (2022-04-06T09:13:25Z) - Decompose to Adapt: Cross-domain Object Detection via Feature
Disentanglement [79.2994130944482]
We design a Domain Disentanglement Faster-RCNN (DDF) to eliminate the source-specific information in the features for detection task learning.
Our DDF method facilitates the feature disentanglement at the global and local stages, with a Global Triplet Disentanglement (GTD) module and an Instance Similarity Disentanglement (ISD) module.
By outperforming state-of-the-art methods on four benchmark UDA object detection tasks, our DDF method is demonstrated to be effective with wide applicability.
arXiv Detail & Related papers (2022-01-06T05:43:01Z) - META: Mimicking Embedding via oThers' Aggregation for Generalizable
Person Re-identification [68.39849081353704]
Domain generalizable (DG) person re-identification (ReID) aims to test across unseen domains without access to the target domain data at training time.
This paper presents a new approach called Mimicking Embedding via oThers' Aggregation (META) for DG ReID.
arXiv Detail & Related papers (2021-12-16T08:06:50Z) - Exploiting Domain-Specific Features to Enhance Domain Generalization [10.774902700296249]
Domain Generalization (DG) aims to train a model, from multiple observed source domains, in order to perform well on unseen target domains.
Prior DG approaches have focused on extracting domain-invariant information across sources to generalize on target domains.
We propose meta-Domain Specific-Domain Invariant (mD) - a novel theoretically sound framework.
arXiv Detail & Related papers (2021-10-18T15:42:39Z) - COLUMBUS: Automated Discovery of New Multi-Level Features for Domain
Generalization via Knowledge Corruption [12.555885317622131]
We address the challenging domain generalization problem, where a model trained on a set of source domains is expected to generalize well in unseen domains without exposure to their data.
We propose Columbus, a method that enforces new feature discovery via a targeted corruption of the most relevant input and multi-level representations of the data.
arXiv Detail & Related papers (2021-09-09T14:52:05Z) - Discriminative Domain-Invariant Adversarial Network for Deep Domain
Generalization [33.84004077585957]
We propose a discriminative domain-invariant adversarial network (DDIAN) for domain generalization.
DDIAN achieves better prediction on unseen target data during training compared to state-of-the-art domain generalization approaches.
arXiv Detail & Related papers (2021-08-20T04:24:12Z) - Discrepancy Minimization in Domain Generalization with Generative
Nearest Neighbors [13.047289562445242]
Domain generalization (DG) deals with the problem of domain shift where a machine learning model trained on multiple-source domains fail to generalize well on a target domain with different statistics.
Multiple approaches have been proposed to solve the problem of domain generalization by learning domain invariant representations across the source domains that fail to guarantee generalization on the shifted target domain.
We propose a Generative Nearest Neighbor based Discrepancy Minimization (GNNDM) method which provides a theoretical guarantee that is upper bounded by the error in the labeling process of the target.
arXiv Detail & Related papers (2020-07-28T14:54:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.