Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment
- URL: http://arxiv.org/abs/2310.13959v2
- Date: Wed, 1 Nov 2023 04:00:51 GMT
- Title: Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment
- Authors: Chuang Zhao, Hongke Zhao, Hengshu Zhu, Zhenya Huang, Nan Feng, Enhong
Chen, Hui Xiong
- Abstract summary: We propose a novel bi-discriminator domain adversarial neural network with class-level gradient alignment.
BACG resorts to gradient signals and second-order probability estimation for better alignment of domain distributions.
In addition, inspired by contrastive learning, we develop a memory bank-based variant, i.e. Fast-BACG, which can greatly shorten the training process.
- Score: 87.8301166955305
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised domain adaptation aims to transfer rich knowledge from the
annotated source domain to the unlabeled target domain with the same label
space. One prevalent solution is the bi-discriminator domain adversarial
network, which strives to identify target domain samples outside the support of
the source domain distribution and enforces their classification to be
consistent on both discriminators. Despite being effective, agnostic accuracy
and overconfident estimation for out-of-distribution samples hinder its further
performance improvement. To address the above challenges, we propose a novel
bi-discriminator domain adversarial neural network with class-level gradient
alignment, i.e. BACG. BACG resorts to gradient signals and second-order
probability estimation for better alignment of domain distributions.
Specifically, for accuracy-awareness, we first design an optimizable nearest
neighbor algorithm to obtain pseudo-labels of samples in the target domain, and
then enforce the backward gradient approximation of the two discriminators at
the class level. Furthermore, following evidential learning theory, we
transform the traditional softmax-based optimization method into a Multinomial
Dirichlet hierarchical model to infer the class probability distribution as
well as samples uncertainty, thereby alleviating misestimation of
out-of-distribution samples and guaranteeing high-quality classes alignment. In
addition, inspired by contrastive learning, we develop a memory bank-based
variant, i.e. Fast-BACG, which can greatly shorten the training process at the
cost of a minor decrease in accuracy. Extensive experiments and detailed
theoretical analysis on four benchmark data sets validate the effectiveness and
robustness of our algorithm.
Related papers
- Adaptive Betweenness Clustering for Semi-Supervised Domain Adaptation [108.40945109477886]
We propose a novel SSDA approach named Graph-based Adaptive Betweenness Clustering (G-ABC) for achieving categorical domain alignment.
Our method outperforms previous state-of-the-art SSDA approaches, demonstrating the superiority of the proposed G-ABC algorithm.
arXiv Detail & Related papers (2024-01-21T09:57:56Z) - Low-confidence Samples Matter for Domain Adaptation [47.552605279925736]
Domain adaptation (DA) aims to transfer knowledge from a label-rich source domain to a related but label-scarce target domain.
We propose a novel contrastive learning method by processing low-confidence samples.
We evaluate the proposed method in both unsupervised and semi-supervised DA settings.
arXiv Detail & Related papers (2022-02-06T15:45:45Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Boosting Unsupervised Domain Adaptation with Soft Pseudo-label and
Curriculum Learning [19.903568227077763]
Unsupervised domain adaptation (UDA) improves classification performance on an unlabeled target domain by leveraging data from a fully labeled source domain.
We propose a model-agnostic two-stage learning framework, which greatly reduces flawed model predictions using soft pseudo-label strategy.
At the second stage, we propose a curriculum learning strategy to adaptively control the weighting between losses from the two domains.
arXiv Detail & Related papers (2021-12-03T14:47:32Z) - Coarse to Fine: Domain Adaptive Crowd Counting via Adversarial Scoring
Network [58.05473757538834]
This paper proposes a novel adversarial scoring network (ASNet) to bridge the gap across domains from coarse to fine granularity.
Three sets of migration experiments show that the proposed methods achieve state-of-the-art counting performance.
arXiv Detail & Related papers (2021-07-27T14:47:24Z) - Instance Level Affinity-Based Transfer for Unsupervised Domain
Adaptation [74.71931918541748]
We propose an instance affinity based criterion for source to target transfer during adaptation, called ILA-DA.
We first propose a reliable and efficient method to extract similar and dissimilar samples across source and target, and utilize a multi-sample contrastive loss to drive the domain alignment process.
We verify the effectiveness of ILA-DA by observing consistent improvements in accuracy over popular domain adaptation approaches on a variety of benchmark datasets.
arXiv Detail & Related papers (2021-04-03T01:33:14Z) - Selective Pseudo-Labeling with Reinforcement Learning for
Semi-Supervised Domain Adaptation [116.48885692054724]
We propose a reinforcement learning based selective pseudo-labeling method for semi-supervised domain adaptation.
We develop a deep Q-learning model to select both accurate and representative pseudo-labeled instances.
Our proposed method is evaluated on several benchmark datasets for SSDA, and demonstrates superior performance to all the comparison methods.
arXiv Detail & Related papers (2020-12-07T03:37:38Z) - Unsupervised Domain Adaptation via Discriminative Manifold Propagation [26.23123292060868]
Unsupervised domain adaptation is effective in leveraging rich information from a labeled source domain to an unlabeled target domain.
The proposed method can be used to tackle a series of variants of domain adaptation problems, including both vanilla and partial settings.
arXiv Detail & Related papers (2020-08-23T12:31:37Z) - Unsupervised Domain Adaptation via Discriminative Manifold Embedding and
Alignment [23.72562139715191]
Unsupervised domain adaptation is effective in leveraging the rich information from the source domain to the unsupervised target domain.
The hard-assigned pseudo labels on the target domain are risky to the intrinsic data structure.
A consistent manifold learning framework is proposed to achieve transferability and discriminability consistently.
arXiv Detail & Related papers (2020-02-20T11:06:41Z) - Unsupervised Domain Adaptive Object Detection using Forward-Backward
Cyclic Adaptation [13.163271874039191]
We present a novel approach to perform the unsupervised domain adaptation for object detection through forward-backward cyclic (FBC) training.
Recent adversarial training based domain adaptation methods have shown their effectiveness on minimizing domain discrepancy via marginal feature distributions alignment.
We propose Forward-Backward Cyclic Adaptation, which iteratively computes adaptation from source to target via backward hopping and from target to source via forward passing.
arXiv Detail & Related papers (2020-02-03T06:24:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.