Domain Adversarial Training: A Game Perspective
- URL: http://arxiv.org/abs/2202.05352v1
- Date: Thu, 10 Feb 2022 22:17:30 GMT
- Title: Domain Adversarial Training: A Game Perspective
- Authors: David Acuna, Marc T Law, Guojun Zhang, Sanja Fidler
- Abstract summary: This paper defines optimal solutions in domain-adversarial training from a game theoretical perspective.
We show that descent in domain-adversarial training can violate the convergence guarantees of the gradient, oftentimes hindering the transfer performance.
Ours are easy to implement, free of additional parameters, and can be plugged into any domain-adversarial framework.
- Score: 80.3821370633883
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The dominant line of work in domain adaptation has focused on learning
invariant representations using domain-adversarial training. In this paper, we
interpret this approach from a game theoretical perspective. Defining optimal
solutions in domain-adversarial training as a local Nash equilibrium, we show
that gradient descent in domain-adversarial training can violate the asymptotic
convergence guarantees of the optimizer, oftentimes hindering the transfer
performance. Our analysis leads us to replace gradient descent with high-order
ODE solvers (i.e., Runge-Kutta), for which we derive asymptotic convergence
guarantees. This family of optimizers is significantly more stable and allows
more aggressive learning rates, leading to high performance gains when used as
a drop-in replacement over standard optimizers. Our experiments show that in
conjunction with state-of-the-art domain-adversarial methods, we achieve up to
3.5% improvement with less than of half training iterations. Our optimizers are
easy to implement, free of additional parameters, and can be plugged into any
domain-adversarial framework.
Related papers
- Improving Instance Optimization in Deformable Image Registration with Gradient Projection [7.6061804149819885]
Deformable image registration is inherently a multi-objective optimization problem.
These conflicting objectives often lead to poor optimization outcomes.
Deep learning methods have recently gained popularity in this domain due to their efficiency in processing large datasets.
arXiv Detail & Related papers (2024-10-21T08:27:13Z) - Domain-Inspired Sharpness-Aware Minimization Under Domain Shifts [38.78485556098491]
This paper presents a Domain-Inspired Sharpness-Aware Minimization (DISAM) algorithm for optimization under domain shifts.
It is motivated by the inconsistent convergence degree of SAM across different domains, which induces optimization bias towards certain domains.
Under this mechanism, we theoretically show that DISAM can achieve faster overall convergence and improved generalization in principle.
arXiv Detail & Related papers (2024-05-29T08:22:33Z) - Adaptive Federated Learning Over the Air [108.62635460744109]
We propose a federated version of adaptive gradient methods, particularly AdaGrad and Adam, within the framework of over-the-air model training.
Our analysis shows that the AdaGrad-based training algorithm converges to a stationary point at the rate of $mathcalO( ln(T) / T 1 - frac1alpha ).
arXiv Detail & Related papers (2024-03-11T09:10:37Z) - A Closer Look at Smoothness in Domain Adversarial Training [37.205372217498656]
We analyze the effect of smoothness enhancing formulations on domain adversarial training.
We find that converging to a smooth minima with respect to (w.r.t.) task loss stabilizes the adversarial training leading to better performance on target domain.
In contrast to task loss, our analysis shows that converging to smooth minima w.r.t. adversarial loss leads to sub-optimal generalization on the target domain.
arXiv Detail & Related papers (2022-06-16T14:31:38Z) - Domain Adaptation for Semantic Segmentation via Patch-Wise Contrastive
Learning [62.7588467386166]
We leverage contrastive learning to bridge the domain gap by aligning the features of structurally similar label patches across domains.
Our approach consistently outperforms state-of-the-art unsupervised and semi-supervised methods on two challenging domain adaptive segmentation tasks.
arXiv Detail & Related papers (2021-04-22T13:39:12Z) - Gradient Matching for Domain Generalization [93.04545793814486]
A critical requirement of machine learning systems is their ability to generalize to unseen domains.
We propose an inter-domain gradient matching objective that targets domain generalization.
We derive a simpler first-order algorithm named Fish that approximates its optimization.
arXiv Detail & Related papers (2021-04-20T12:55:37Z) - Latent-Optimized Adversarial Neural Transfer for Sarcasm Detection [50.29565896287595]
We apply transfer learning to exploit common datasets for sarcasm detection.
We propose a generalized latent optimization strategy that allows different losses to accommodate each other.
In particular, we achieve 10.02% absolute performance gain over the previous state of the art on the iSarcasm dataset.
arXiv Detail & Related papers (2021-04-19T13:07:52Z) - Pixel-Level Cycle Association: A New Perspective for Domain Adaptive
Semantic Segmentation [169.82760468633236]
We propose to build the pixel-level cycle association between source and target pixel pairs.
Our method can be trained end-to-end in one stage and introduces no additional parameters.
arXiv Detail & Related papers (2020-10-31T00:11:36Z) - Unsupervised Domain Adaptive Object Detection using Forward-Backward
Cyclic Adaptation [13.163271874039191]
We present a novel approach to perform the unsupervised domain adaptation for object detection through forward-backward cyclic (FBC) training.
Recent adversarial training based domain adaptation methods have shown their effectiveness on minimizing domain discrepancy via marginal feature distributions alignment.
We propose Forward-Backward Cyclic Adaptation, which iteratively computes adaptation from source to target via backward hopping and from target to source via forward passing.
arXiv Detail & Related papers (2020-02-03T06:24:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.