Learning Generalizable Models via Disentangling Spurious and Enhancing
Potential Correlations
- URL: http://arxiv.org/abs/2401.05752v1
- Date: Thu, 11 Jan 2024 09:00:22 GMT
- Title: Learning Generalizable Models via Disentangling Spurious and Enhancing
Potential Correlations
- Authors: Na Wang, Lei Qi, Jintao Guo, Yinghuan Shi, Yang Gao
- Abstract summary: Domain generalization (DG) aims to train a model on multiple source domains to ensure that it can generalize well to an arbitrary unseen target domain.
Adopting multiple perspectives, such as the sample and the feature, proves to be effective.
In this paper, we focus on improving the generalization ability of the model by compelling it to acquire domain-invariant representations from both the sample and feature perspectives.
- Score: 28.38895118573957
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Domain generalization (DG) intends to train a model on multiple source
domains to ensure that it can generalize well to an arbitrary unseen target
domain. The acquisition of domain-invariant representations is pivotal for DG
as they possess the ability to capture the inherent semantic information of the
data, mitigate the influence of domain shift, and enhance the generalization
capability of the model. Adopting multiple perspectives, such as the sample and
the feature, proves to be effective. The sample perspective facilitates data
augmentation through data manipulation techniques, whereas the feature
perspective enables the extraction of meaningful generalization features. In
this paper, we focus on improving the generalization ability of the model by
compelling it to acquire domain-invariant representations from both the sample
and feature perspectives by disentangling spurious correlations and enhancing
potential correlations. 1) From the sample perspective, we develop a frequency
restriction module, guiding the model to focus on the relevant correlations
between object features and labels, thereby disentangling spurious
correlations. 2) From the feature perspective, the simple Tail Interaction
module implicitly enhances potential correlations among all samples from all
source domains, facilitating the acquisition of domain-invariant
representations across multiple domains for the model. The experimental results
show that Convolutional Neural Networks (CNNs) or Multi-Layer Perceptrons
(MLPs) with a strong baseline embedded with these two modules can achieve
superior results, e.g., an average accuracy of 92.30% on Digits-DG.
Related papers
- Dual-stream Feature Augmentation for Domain Generalization [16.495752769624872]
We propose a Dual-stream Feature Augmentation(DFA) method by constructing some hard features from two perspectives.
Our approach could achieve state-of-the-art performance for domain generalization.
arXiv Detail & Related papers (2024-09-07T03:41:05Z) - A Plug-and-Play Method for Rare Human-Object Interactions Detection by Bridging Domain Gap [50.079224604394]
We present a novel model-agnostic framework called textbfContext-textbfEnhanced textbfFeature textbfAment (CEFA)
CEFA consists of a feature alignment module and a context enhancement module.
Our method can serve as a plug-and-play module to improve the detection performance of HOI models on rare categories.
arXiv Detail & Related papers (2024-07-31T08:42:48Z) - Graph-based Unsupervised Disentangled Representation Learning via Multimodal Large Language Models [42.17166746027585]
We introduce a bidirectional weighted graph-based framework to learn factorized attributes and their interrelations within complex data.
Specifically, we propose a $beta$-VAE based module to extract factors as the initial nodes of the graph.
By integrating these complementary modules, our model successfully achieves fine-grained, practical and unsupervised disentanglement.
arXiv Detail & Related papers (2024-07-26T15:32:21Z) - Learning Divergence Fields for Shift-Robust Graph Representations [73.11818515795761]
In this work, we propose a geometric diffusion model with learnable divergence fields for the challenging problem with interdependent data.
We derive a new learning objective through causal inference, which can guide the model to learn generalizable patterns of interdependence that are insensitive across domains.
arXiv Detail & Related papers (2024-06-07T14:29:21Z) - Learning to Learn Domain-invariant Parameters for Domain Generalization [29.821634033299855]
Domain generalization (DG) aims to overcome this issue by capturing domain-invariant representations from source domains.
We propose two modules of Domain Decoupling and Combination (DDC) and Domain-invariance-guided Backpropagation (DIGB)
Our proposed method has achieved state-of-the-art performance with strong generalization capability.
arXiv Detail & Related papers (2022-11-04T07:19:34Z) - Learning to Augment via Implicit Differentiation for Domain
Generalization [107.9666735637355]
Domain generalization (DG) aims to overcome the problem by leveraging multiple source domains to learn a domain-generalizable model.
In this paper, we propose a novel augmentation-based DG approach, dubbed AugLearn.
AugLearn shows effectiveness on three standard DG benchmarks, PACS, Office-Home and Digits-DG.
arXiv Detail & Related papers (2022-10-25T18:51:51Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z) - Instrumental Variable-Driven Domain Generalization with Unobserved
Confounders [53.735614014067394]
Domain generalization (DG) aims to learn from multiple source domains a model that can generalize well on unseen target domains.
We propose an instrumental variable-driven DG method (IV-DG) by removing the bias of the unobserved confounders with two-stage learning.
In the first stage, it learns the conditional distribution of the input features of one domain given input features of another domain.
In the second stage, it estimates the relationship by predicting labels with the learned conditional distribution.
arXiv Detail & Related papers (2021-10-04T13:32:57Z) - Bayesian Sparse Factor Analysis with Kernelized Observations [67.60224656603823]
Multi-view problems can be faced with latent variable models.
High-dimensionality and non-linear issues are traditionally handled by kernel methods.
We propose merging both approaches into single model.
arXiv Detail & Related papers (2020-06-01T14:25:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.