DGFamba: Learning Flow Factorized State Space for Visual Domain Generalization
- URL: http://arxiv.org/abs/2504.08019v1
- Date: Thu, 10 Apr 2025 17:24:53 GMT
- Title: DGFamba: Learning Flow Factorized State Space for Visual Domain Generalization
- Authors: Qi Bi, Jingjun Yi, Hao Zheng, Haolan Zhan, Wei Ji, Yawen Huang, Yuexiang Li,
- Abstract summary: We propose a novel Flow Factorized State Space model, dubbed as DG-Famba, for visual domain generalization.<n>To maintain domain consistency, we innovatively map the style-augmented and the original state embeddings by flow factorization.<n>Experiments conducted on various visual domain generalization settings show its state-of-the-art performance.
- Score: 27.903842187045118
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Domain generalization aims to learn a representation from the source domain, which can be generalized to arbitrary unseen target domains. A fundamental challenge for visual domain generalization is the domain gap caused by the dramatic style variation whereas the image content is stable. The realm of selective state space, exemplified by VMamba, demonstrates its global receptive field in representing the content. However, the way exploiting the domain-invariant property for selective state space is rarely explored. In this paper, we propose a novel Flow Factorized State Space model, dubbed as DG-Famba, for visual domain generalization. To maintain domain consistency, we innovatively map the style-augmented and the original state embeddings by flow factorization. In this latent flow space, each state embedding from a certain style is specified by a latent probability path. By aligning these probability paths in the latent space, the state embeddings are able to represent the same content distribution regardless of the style differences. Extensive experiments conducted on various visual domain generalization settings show its state-of-the-art performance.
Related papers
- StyDeSty: Min-Max Stylization and Destylization for Single Domain Generalization [85.18995948334592]
Single domain generalization (single DG) aims at learning a robust model generalizable to unseen domains from only one training domain.
State-of-the-art approaches have mostly relied on data augmentations, such as adversarial perturbation and style enhancement, to synthesize new data.
We propose emphStyDeSty, which explicitly accounts for the alignment of the source and pseudo domains in the process of data augmentation.
arXiv Detail & Related papers (2024-06-01T02:41:34Z) - DomainDrop: Suppressing Domain-Sensitive Channels for Domain
Generalization [25.940491294232956]
DomainDrop is a framework to continuously enhance the channel robustness to domain shifts.
Our framework achieves state-of-the-art performance compared to other competing methods.
arXiv Detail & Related papers (2023-08-20T14:48:52Z) - Single Domain Dynamic Generalization for Iris Presentation Attack
Detection [41.126916126040655]
Iris presentation generalization has achieved great success under intra-domain settings but easily degrades on unseen domains.
We propose a Single Domain Dynamic Generalization (SDDG) framework, which exploits domain-invariant and domain-specific features on a per-sample basis.
The proposed method is effective and outperforms the state-of-the-art on LivDet-Iris 2017 dataset.
arXiv Detail & Related papers (2023-05-22T07:54:13Z) - Instance-Aware Domain Generalization for Face Anti-Spoofing [42.36157210235893]
Face anti-spoofing (FAS) has been recently studied to improve the generalization on unseen scenarios.
Previous methods rely on domain labels to align the distribution of each domain for learning domain-invariant representations.
We propose a novel perspective for DG FAS that aligns features on the instance level without the need for domain labels.
arXiv Detail & Related papers (2023-04-12T06:41:16Z) - Rethinking Domain Generalization for Face Anti-spoofing: Separability
and Alignment [35.67771212285966]
This work studies the generalization issue of face anti-spoofing (FAS) models on domain gaps, such as image resolution, blurriness and sensor variations.
We formulate this FAS strategy of separability and alignment (SA-FAS) as a problem of invariant risk minimization (IRM)
We demonstrate the effectiveness of SA-FAS on challenging cross-domain FAS datasets and establish state-of-the-art performance.
arXiv Detail & Related papers (2023-03-23T20:34:27Z) - Normalization Perturbation: A Simple Domain Generalization Method for
Real-World Domain Shifts [133.99270341855728]
Real-world domain styles can vary substantially due to environment changes and sensor noises.
Deep models only know the training domain style.
We propose Normalization Perturbation to overcome this domain style overfitting problem.
arXiv Detail & Related papers (2022-11-08T17:36:49Z) - Semantic-Aware Domain Generalized Segmentation [67.49163582961877]
Deep models trained on source domain lack generalization when evaluated on unseen target domains with different data distributions.
We propose a framework including two novel modules: Semantic-Aware Normalization (SAN) and Semantic-Aware Whitening (SAW)
Our approach shows significant improvements over existing state-of-the-art on various backbone networks.
arXiv Detail & Related papers (2022-04-02T09:09:59Z) - Domain Generalization via Shuffled Style Assembly for Face Anti-Spoofing [69.80851569594924]
Generalizable face anti-spoofing (FAS) has drawn growing attention.
In this work, we separate the complete representation into content and style ones.
A novel Shuffled Style Assembly Network (SSAN) is proposed to extract and reassemble different content and style features.
arXiv Detail & Related papers (2022-03-10T12:44:05Z) - Structured Latent Embeddings for Recognizing Unseen Classes in Unseen
Domains [108.11746235308046]
We propose a novel approach that learns domain-agnostic structured latent embeddings by projecting images from different domains.
Our experiments on the challenging DomainNet and DomainNet-LS benchmarks show the superiority of our approach over existing methods.
arXiv Detail & Related papers (2021-07-12T17:57:46Z) - Batch Normalization Embeddings for Deep Domain Generalization [50.51405390150066]
Domain generalization aims at training machine learning models to perform robustly across different and unseen domains.
We show a significant increase in classification accuracy over current state-of-the-art techniques on popular domain generalization benchmarks.
arXiv Detail & Related papers (2020-11-25T12:02:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.