f-Domain-Adversarial Learning: Theory and Algorithms
- URL: http://arxiv.org/abs/2106.11344v1
- Date: Mon, 21 Jun 2021 18:21:09 GMT
- Title: f-Domain-Adversarial Learning: Theory and Algorithms
- Authors: David Acuna, Guojun Zhang, Marc T. Law, Sanja Fidler
- Abstract summary: Unsupervised domain adaptation is used in many machine learning applications where, during training, a model has access to unlabeled data in the target domain.
We derive a novel generalization bound for domain adaptation that exploits a new measure of discrepancy between distributions based on a variational characterization of f-divergences.
- Score: 82.97698406515667
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised domain adaptation is used in many machine learning applications
where, during training, a model has access to unlabeled data in the target
domain, and a related labeled dataset. In this paper, we introduce a novel and
general domain-adversarial framework. Specifically, we derive a novel
generalization bound for domain adaptation that exploits a new measure of
discrepancy between distributions based on a variational characterization of
f-divergences. It recovers the theoretical results from Ben-David et al.
(2010a) as a special case and supports divergences used in practice. Based on
this bound, we derive a new algorithmic framework that introduces a key
correction in the original adversarial training method of Ganin et al. (2016).
We show that many regularizers and ad-hoc objectives introduced over the last
years in this framework are then not required to achieve performance comparable
to (if not better than) state-of-the-art domain-adversarial methods.
Experimental analysis conducted on real-world natural language and computer
vision datasets show that our framework outperforms existing baselines, and
obtains the best results for f-divergences that were not considered previously
in domain-adversarial learning.
Related papers
- Understanding the Cross-Domain Capabilities of Video-Based Few-Shot Action Recognition Models [3.072340427031969]
Few-shot action recognition (FSAR) aims to learn a model capable of identifying novel actions in videos using only a few examples.
In assuming the base dataset seen during meta-training and novel dataset used for evaluation can come from different domains, cross-domain few-shot learning alleviates data collection and annotation costs.
We systematically evaluate existing state-of-the-art single-domain, transfer-based, and cross-domain FSAR methods on new cross-domain tasks.
arXiv Detail & Related papers (2024-06-03T07:48:18Z) - Non-stationary Domain Generalization: Theory and Algorithm [11.781050299571692]
In this paper, we study domain generalization in non-stationary environment.
We first examine the impact of environmental non-stationarity on model performance.
Then, we propose a novel algorithm based on adaptive invariant representation learning.
arXiv Detail & Related papers (2024-05-10T21:32:43Z) - Federated Domain Generalization: A Survey [12.84261944926547]
In machine learning, data is often distributed across different devices, organizations, or edge nodes.
In response to this challenge, there has been a surge of interest in federated domain generalization.
This paper presents the first survey of recent advances in this area.
arXiv Detail & Related papers (2023-06-02T07:55:42Z) - Meta-causal Learning for Single Domain Generalization [102.53303707563612]
Single domain generalization aims to learn a model from a single training domain (source domain) and apply it to multiple unseen test domains (target domains)
Existing methods focus on expanding the distribution of the training domain to cover the target domains, but without estimating the domain shift between the source and target domains.
We propose a new learning paradigm, namely simulate-analyze-reduce, which first simulates the domain shift by building an auxiliary domain as the target domain, then learns to analyze the causes of domain shift, and finally learns to reduce the domain shift for model adaptation.
arXiv Detail & Related papers (2023-04-07T15:46:38Z) - Improving Domain Generalization with Domain Relations [77.63345406973097]
This paper focuses on domain shifts, which occur when the model is applied to new domains that are different from the ones it was trained on.
We propose a new approach called D$3$G to learn domain-specific models.
Our results show that D$3$G consistently outperforms state-of-the-art methods.
arXiv Detail & Related papers (2023-02-06T08:11:16Z) - Self-balanced Learning For Domain Generalization [64.99791119112503]
Domain generalization aims to learn a prediction model on multi-domain source data such that the model can generalize to a target domain with unknown statistics.
Most existing approaches have been developed under the assumption that the source data is well-balanced in terms of both domain and class.
We propose a self-balanced domain generalization framework that adaptively learns the weights of losses to alleviate the bias caused by different distributions of the multi-domain source data.
arXiv Detail & Related papers (2021-08-31T03:17:54Z) - Model-Based Domain Generalization [96.84818110323518]
We propose a novel approach for the domain generalization problem called Model-Based Domain Generalization.
Our algorithms beat the current state-of-the-art methods on the very-recently-proposed WILDS benchmark by up to 20 percentage points.
arXiv Detail & Related papers (2021-02-23T00:59:02Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.