CAusal and collaborative proxy-tasKs lEarning for Semi-Supervised Domain
Adaptation
- URL: http://arxiv.org/abs/2303.17526v1
- Date: Thu, 30 Mar 2023 16:48:28 GMT
- Title: CAusal and collaborative proxy-tasKs lEarning for Semi-Supervised Domain
Adaptation
- Authors: Wenqiao Zhang, Changshuo Liu, Can Cui, Beng Chin Ooi
- Abstract summary: Semi-supervised domain adaptation (SSDA) adapts a learner to a new domain by effectively utilizing source domain data and a few labeled target samples.
We show that the proposed model significantly outperforms SOTA methods in terms of effectiveness and generalisability on SSDA datasets.
- Score: 20.589323508870592
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semi-supervised domain adaptation (SSDA) adapts a learner to a new domain by
effectively utilizing source domain data and a few labeled target samples. It
is a practical yet under-investigated research topic. In this paper, we analyze
the SSDA problem from two perspectives that have previously been overlooked,
and correspondingly decompose it into two \emph{key subproblems}: \emph{robust
domain adaptation (DA) learning} and \emph{maximal cross-domain data
utilization}. \textbf{(i)} From a causal theoretical view, a robust DA model
should distinguish the invariant ``concept'' (key clue to image label) from the
nuisance of confounding factors across domains. To achieve this goal, we
propose to generate \emph{concept-invariant samples} to enable the model to
classify the samples through causal intervention, yielding improved
generalization guarantees; \textbf{(ii)} Based on the robust DA theory, we aim
to exploit the maximal utilization of rich source domain data and a few labeled
target samples to boost SSDA further. Consequently, we propose a
collaboratively debiasing learning framework that utilizes two complementary
semi-supervised learning (SSL) classifiers to mutually exchange their unbiased
knowledge, which helps unleash the potential of source and target domain
training data, thereby producing more convincing pseudo-labels. Such obtained
labels facilitate cross-domain feature alignment and duly improve the invariant
concept learning. In our experimental study, we show that the proposed model
significantly outperforms SOTA methods in terms of effectiveness and
generalisability on SSDA datasets.
Related papers
- High-order Neighborhoods Know More: HyperGraph Learning Meets Source-free Unsupervised Domain Adaptation [34.08681468394247]
Source-free Unsupervised Domain Adaptation aims to classify target samples by only accessing a pre-trained source model and unlabelled target samples.
Existing methods normally exploit the pair-wise relation among target samples and attempt to discover their correlations by clustering these samples based on semantic features.
We propose a new SFDA method that exploits the high-order neighborhood relation and explicitly takes the domain shift effect into account.
arXiv Detail & Related papers (2024-05-11T05:07:43Z) - A Robust Negative Learning Approach to Partial Domain Adaptation Using
Source Prototypes [0.8895157045883034]
This work proposes a robust Partial Domain Adaptation (PDA) framework that mitigates the negative transfer problem.
It includes diverse, complementary label feedback, alleviating the effect of incorrect feedback and promoting pseudo-label refinement.
We conducted a series of comprehensive experiments, including an ablation analysis, covering a range of partial domain adaptation tasks.
arXiv Detail & Related papers (2023-09-07T07:26:27Z) - Divide and Contrast: Source-free Domain Adaptation via Adaptive
Contrastive Learning [122.62311703151215]
Divide and Contrast (DaC) aims to connect the good ends of both worlds while bypassing their limitations.
DaC divides the target data into source-like and target-specific samples, where either group of samples is treated with tailored goals.
We further align the source-like domain with the target-specific samples using a memory bank-based Maximum Mean Discrepancy (MMD) loss to reduce the distribution mismatch.
arXiv Detail & Related papers (2022-11-12T09:21:49Z) - Source-Free Domain Adaptation via Distribution Estimation [106.48277721860036]
Domain Adaptation aims to transfer the knowledge learned from a labeled source domain to an unlabeled target domain whose data distributions are different.
Recently, Source-Free Domain Adaptation (SFDA) has drawn much attention, which tries to tackle domain adaptation problem without using source data.
In this work, we propose a novel framework called SFDA-DE to address SFDA task via source Distribution Estimation.
arXiv Detail & Related papers (2022-04-24T12:22:19Z) - Low-confidence Samples Matter for Domain Adaptation [47.552605279925736]
Domain adaptation (DA) aims to transfer knowledge from a label-rich source domain to a related but label-scarce target domain.
We propose a novel contrastive learning method by processing low-confidence samples.
We evaluate the proposed method in both unsupervised and semi-supervised DA settings.
arXiv Detail & Related papers (2022-02-06T15:45:45Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Learning Invariant Representation with Consistency and Diversity for
Semi-supervised Source Hypothesis Transfer [46.68586555288172]
We propose a novel task named Semi-supervised Source Hypothesis Transfer (SSHT), which performs domain adaptation based on source trained model, to generalize well in target domain with a few supervisions.
We propose Consistency and Diversity Learning (CDL), a simple but effective framework for SSHT by facilitating prediction consistency between two randomly augmented unlabeled data.
Experimental results show that our method outperforms existing SSDA methods and unsupervised model adaptation methods on DomainNet, Office-Home and Office-31 datasets.
arXiv Detail & Related papers (2021-07-07T04:14:24Z) - Source-Free Open Compound Domain Adaptation in Semantic Segmentation [99.82890571842603]
In SF-OCDA, only the source pre-trained model and the target data are available to learn the target model.
We propose the Cross-Patch Style Swap (CPSS) to diversify samples with various patch styles in the feature-level.
Our method produces state-of-the-art results on the C-Driving dataset.
arXiv Detail & Related papers (2021-06-07T08:38:41Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z) - Universal Source-Free Domain Adaptation [57.37520645827318]
We propose a novel two-stage learning process for domain adaptation.
In the Procurement stage, we aim to equip the model for future source-free deployment, assuming no prior knowledge of the upcoming category-gap and domain-shift.
In the Deployment stage, the goal is to design a unified adaptation algorithm capable of operating across a wide range of category-gaps.
arXiv Detail & Related papers (2020-04-09T07:26:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.