Towards Realizing the Value of Labeled Target Samples: a Two-Stage
Approach for Semi-Supervised Domain Adaptation
- URL: http://arxiv.org/abs/2304.10762v1
- Date: Fri, 21 Apr 2023 06:13:23 GMT
- Title: Towards Realizing the Value of Labeled Target Samples: a Two-Stage
Approach for Semi-Supervised Domain Adaptation
- Authors: mengqun Jin, Kai Li, Shuyan Li, Chunming He, Xiu Li
- Abstract summary: Semi-Supervised Domain Adaptation (SSDA) is a recently emerging research topic that extends from the widely-investigated Unsupervised Domain Adaptation (UDA)
We propose to decouple SSDA as an UDA problem and a semi-supervised learning problem where we first learn an UDA model using labeled source and unlabeled target samples.
By utilizing the labeled source samples and target samples separately, the bias problem can be well mitigated.
- Score: 21.004477294264998
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semi-Supervised Domain Adaptation (SSDA) is a recently emerging research
topic that extends from the widely-investigated Unsupervised Domain Adaptation
(UDA) by further having a few target samples labeled, i.e., the model is
trained with labeled source samples, unlabeled target samples as well as a few
labeled target samples. Compared with UDA, the key to SSDA lies how to most
effectively utilize the few labeled target samples. Existing SSDA approaches
simply merge the few precious labeled target samples into vast labeled source
samples or further align them, which dilutes the value of labeled target
samples and thus still obtains a biased model. To remedy this, in this paper,
we propose to decouple SSDA as an UDA problem and a semi-supervised learning
problem where we first learn an UDA model using labeled source and unlabeled
target samples and then adapt the learned UDA model in a semi-supervised way
using labeled and unlabeled target samples. By utilizing the labeled source
samples and target samples separately, the bias problem can be well mitigated.
We further propose a consistency learning based mean teacher model to
effectively adapt the learned UDA model using labeled and unlabeled target
samples. Experiments show our approach outperforms existing methods.
Related papers
- Learning from Different Samples: A Source-free Framework for Semi-supervised Domain Adaptation [20.172605920901777]
This paper focuses on designing a framework to use different strategies for comprehensively mining different target samples.
We propose a novel source-free framework (SOUF) to achieve semi-supervised fine-tuning of the source pre-trained model on the target domain.
arXiv Detail & Related papers (2024-11-11T02:09:32Z) - Not All Samples Should Be Utilized Equally: Towards Understanding and Improving Dataset Distillation [57.6797306341115]
We take an initial step towards understanding various matching-based DD methods from the perspective of sample difficulty.
We then extend the neural scaling laws of data pruning to DD to theoretically explain these matching-based methods.
We introduce the Sample Difficulty Correction (SDC) approach, designed to predominantly generate easier samples to achieve higher dataset quality.
arXiv Detail & Related papers (2024-08-22T15:20:32Z) - Adversarial Semi-Supervised Domain Adaptation for Semantic Segmentation:
A New Role for Labeled Target Samples [7.199108088621308]
We design new training objective losses for cases when labeled target data behave as source samples or as real target samples.
To support our approach, we consider a complementary method that mixes source and labeled target data, then applies the same adaptation process.
We illustrate our findings through extensive experiments on the benchmarks GTA5, SYNTHIA, and Cityscapes.
arXiv Detail & Related papers (2023-12-12T15:40:22Z) - How Low Can You Go? Surfacing Prototypical In-Distribution Samples for Unsupervised Anomaly Detection [48.30283806131551]
We show that UAD with extremely few training samples can already match -- and in some cases even surpass -- the performance of training with the whole training dataset.
We propose an unsupervised method to reliably identify prototypical samples to further boost UAD performance.
arXiv Detail & Related papers (2023-12-06T15:30:47Z) - Divide and Contrast: Source-free Domain Adaptation via Adaptive
Contrastive Learning [122.62311703151215]
Divide and Contrast (DaC) aims to connect the good ends of both worlds while bypassing their limitations.
DaC divides the target data into source-like and target-specific samples, where either group of samples is treated with tailored goals.
We further align the source-like domain with the target-specific samples using a memory bank-based Maximum Mean Discrepancy (MMD) loss to reduce the distribution mismatch.
arXiv Detail & Related papers (2022-11-12T09:21:49Z) - ReSmooth: Detecting and Utilizing OOD Samples when Training with Data
Augmentation [57.38418881020046]
Recent DA techniques always meet the need for diversity in augmented training samples.
An augmentation strategy that has a high diversity usually introduces out-of-distribution (OOD) augmented samples.
We propose ReSmooth, a framework that firstly detects OOD samples in augmented samples and then leverages them.
arXiv Detail & Related papers (2022-05-25T09:29:27Z) - UMAD: Universal Model Adaptation under Domain and Category Shift [138.12678159620248]
Universal Model ADaptation (UMAD) framework handles both UDA scenarios without access to source data.
We develop an informative consistency score to help distinguish unknown samples from known samples.
Experiments on open-set and open-partial-set UDA scenarios demonstrate that UMAD exhibits comparable, if not superior, performance to state-of-the-art data-dependent methods.
arXiv Detail & Related papers (2021-12-16T01:22:59Z) - Semi-supervised Domain Adaptation via Sample-to-Sample Self-Distillation [30.556465364799948]
Semi-supervised domain adaptation (SSDA) is to adapt a learner to a new domain with only a small set of labeled samples when a large labeled dataset is given on a source domain.
We propose a pair-based SSDA method that adapts a model to the target domain using self-distillation with sample pairs.
arXiv Detail & Related papers (2021-11-29T07:13:36Z) - Cross-Domain Gradient Discrepancy Minimization for Unsupervised Domain
Adaptation [22.852237073492894]
Unsupervised Domain Adaptation (UDA) aims to generalize the knowledge learned from a well-labeled source domain to an unlabeled target domain.
We propose a cross-domain discrepancy minimization (CGDM) method which explicitly minimizes the discrepancy of gradients generated by source samples and target samples.
In order to compute the gradient signal of target samples, we further obtain target pseudo labels through a clustering-based self-supervised learning.
arXiv Detail & Related papers (2021-06-08T07:35:40Z) - Semi-Supervised Domain Adaptation with Prototypical Alignment and
Consistency Learning [86.6929930921905]
This paper studies how much it can help address domain shifts if we further have a few target samples labeled.
To explore the full potential of landmarks, we incorporate a prototypical alignment (PA) module which calculates a target prototype for each class from the landmarks.
Specifically, we severely perturb the labeled images, making PA non-trivial to achieve and thus promoting model generalizability.
arXiv Detail & Related papers (2021-04-19T08:46:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.