Tune it the Right Way: Unsupervised Validation of Domain Adaptation via
Soft Neighborhood Density
- URL: http://arxiv.org/abs/2108.10860v1
- Date: Tue, 24 Aug 2021 17:41:45 GMT
- Title: Tune it the Right Way: Unsupervised Validation of Domain Adaptation via
Soft Neighborhood Density
- Authors: Kuniaki Saito, Donghyun Kim, Piotr Teterwak, Stan Sclaroff, Trevor
Darrell, and Kate Saenko
- Abstract summary: We propose an unsupervised validation criterion that measures the density of soft neighborhoods by computing the entropy of the similarity distribution between points.
Our criterion is simpler than competing validation methods, yet more effective.
- Score: 125.64297244986552
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised domain adaptation (UDA) methods can dramatically improve
generalization on unlabeled target domains. However, optimal hyper-parameter
selection is critical to achieving high accuracy and avoiding negative
transfer. Supervised hyper-parameter validation is not possible without labeled
target data, which raises the question: How can we validate unsupervised
adaptation techniques in a realistic way? We first empirically analyze existing
criteria and demonstrate that they are not very effective for tuning
hyper-parameters. Intuitively, a well-trained source classifier should embed
target samples of the same class nearby, forming dense neighborhoods in feature
space. Based on this assumption, we propose a novel unsupervised validation
criterion that measures the density of soft neighborhoods by computing the
entropy of the similarity distribution between points. Our criterion is simpler
than competing validation methods, yet more effective; it can tune
hyper-parameters and the number of training iterations in both image
classification and semantic segmentation models. The code used for the paper
will be available at \url{https://github.com/VisionLearningGroup/SND}.
Related papers
- Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment [87.8301166955305]
We propose a novel bi-discriminator domain adversarial neural network with class-level gradient alignment.
BACG resorts to gradient signals and second-order probability estimation for better alignment of domain distributions.
In addition, inspired by contrastive learning, we develop a memory bank-based variant, i.e. Fast-BACG, which can greatly shorten the training process.
arXiv Detail & Related papers (2023-10-21T09:53:17Z) - Better Practices for Domain Adaptation [62.70267990659201]
Domain adaptation (DA) aims to provide frameworks for adapting models to deployment data without using labels.
Unclear validation protocol for DA has led to bad practices in the literature.
We show challenges across all three branches of domain adaptation methodology.
arXiv Detail & Related papers (2023-09-07T17:44:18Z) - A Study of Unsupervised Evaluation Metrics for Practical and Automatic
Domain Adaptation [15.728090002818963]
Unsupervised domain adaptation (UDA) methods facilitate the transfer of models to target domains without labels.
In this paper, we aim to find an evaluation metric capable of assessing the quality of a transferred model without access to target validation labels.
arXiv Detail & Related papers (2023-08-01T05:01:05Z) - Digging Into Uncertainty-based Pseudo-label for Robust Stereo Matching [39.959000340261625]
We propose to dig into uncertainty estimation for robust stereo matching.
An uncertainty-based pseudo-label is proposed to adapt the pre-trained model to the new domain.
Our method shows strong cross-domain, adapt, and joint generalization and obtains textbf1st place on the stereo task of Robust Vision Challenge 2020.
arXiv Detail & Related papers (2023-07-31T09:11:31Z) - Dual Moving Average Pseudo-Labeling for Source-Free Inductive Domain
Adaptation [45.024029784248825]
Unsupervised domain adaptation reduces the reliance on data annotation in deep learning by adapting knowledge from a source to a target domain.
For privacy and efficiency concerns, source-free domain adaptation extends unsupervised domain adaptation by adapting a pre-trained source model to an unlabeled target domain.
We propose a new semi-supervised fine-tuning method named Dual Moving Average Pseudo-Labeling (DMAPL) for source-free inductive domain adaptation.
arXiv Detail & Related papers (2022-12-15T23:20:13Z) - Divide and Contrast: Source-free Domain Adaptation via Adaptive
Contrastive Learning [122.62311703151215]
Divide and Contrast (DaC) aims to connect the good ends of both worlds while bypassing their limitations.
DaC divides the target data into source-like and target-specific samples, where either group of samples is treated with tailored goals.
We further align the source-like domain with the target-specific samples using a memory bank-based Maximum Mean Discrepancy (MMD) loss to reduce the distribution mismatch.
arXiv Detail & Related papers (2022-11-12T09:21:49Z) - Robust Target Training for Multi-Source Domain Adaptation [110.77704026569499]
We propose a novel Bi-level Optimization based Robust Target Training (BORT$2$) method for MSDA.
Our proposed method achieves the state of the art performance on three MSDA benchmarks, including the large-scale DomainNet dataset.
arXiv Detail & Related papers (2022-10-04T15:20:01Z) - Low-confidence Samples Matter for Domain Adaptation [47.552605279925736]
Domain adaptation (DA) aims to transfer knowledge from a label-rich source domain to a related but label-scarce target domain.
We propose a novel contrastive learning method by processing low-confidence samples.
We evaluate the proposed method in both unsupervised and semi-supervised DA settings.
arXiv Detail & Related papers (2022-02-06T15:45:45Z) - On Universal Black-Box Domain Adaptation [53.7611757926922]
We study an arguably least restrictive setting of domain adaptation in a sense of practical deployment.
Only the interface of source model is available to the target domain, and where the label-space relations between the two domains are allowed to be different and unknown.
We propose to unify them into a self-training framework, regularized by consistency of predictions in local neighborhoods of target samples.
arXiv Detail & Related papers (2021-04-10T02:21:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.