Unsupervised Domain Adaptation via Distilled Discriminative Clustering
- URL: http://arxiv.org/abs/2302.11984v1
- Date: Thu, 23 Feb 2023 13:03:48 GMT
- Title: Unsupervised Domain Adaptation via Distilled Discriminative Clustering
- Authors: Hui Tang, Yaowei Wang, and Kui Jia
- Abstract summary: We re-cast the domain adaptation problem as discriminative clustering of target data.
We propose to jointly train the network using parallel, supervised learning objectives over labeled source data.
We conduct careful ablation studies and extensive experiments on five popular benchmark datasets.
- Score: 45.39542287480395
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised domain adaptation addresses the problem of classifying data in
an unlabeled target domain, given labeled source domain data that share a
common label space but follow a different distribution. Most of the recent
methods take the approach of explicitly aligning feature distributions between
the two domains. Differently, motivated by the fundamental assumption for
domain adaptability, we re-cast the domain adaptation problem as discriminative
clustering of target data, given strong privileged information provided by the
closely related, labeled source data. Technically, we use clustering objectives
based on a robust variant of entropy minimization that adaptively filters
target data, a soft Fisher-like criterion, and additionally the cluster
ordering via centroid classification. To distill discriminative source
information for target clustering, we propose to jointly train the network
using parallel, supervised learning objectives over labeled source data. We
term our method of distilled discriminative clustering for domain adaptation as
DisClusterDA. We also give geometric intuition that illustrates how constituent
objectives of DisClusterDA help learn class-wisely pure, compact feature
distributions. We conduct careful ablation studies and extensive experiments on
five popular benchmark datasets, including a multi-source domain adaptation
one. Based on commonly used backbone networks, DisClusterDA outperforms
existing methods on these benchmarks. It is also interesting to observe that in
our DisClusterDA framework, adding an additional loss term that explicitly
learns to align class-level feature distributions across domains does harm to
the adaptation performance, though more careful studies in different
algorithmic frameworks are to be conducted.
Related papers
- Polycentric Clustering and Structural Regularization for Source-free
Unsupervised Domain Adaptation [20.952542421577487]
Source-Free Domain Adaptation (SFDA) aims to solve the domain adaptation problem by transferring the knowledge learned from a pre-trained source model to an unseen target domain.
Most existing methods assign pseudo-labels to the target data by generating feature prototypes.
In this paper, a novel framework named PCSR is proposed to tackle SFDA via a novel intra-class Polycentric Clustering and Structural Regularization strategy.
arXiv Detail & Related papers (2022-10-14T02:20:48Z) - Source-Free Domain Adaptation via Distribution Estimation [106.48277721860036]
Domain Adaptation aims to transfer the knowledge learned from a labeled source domain to an unlabeled target domain whose data distributions are different.
Recently, Source-Free Domain Adaptation (SFDA) has drawn much attention, which tries to tackle domain adaptation problem without using source data.
In this work, we propose a novel framework called SFDA-DE to address SFDA task via source Distribution Estimation.
arXiv Detail & Related papers (2022-04-24T12:22:19Z) - Cross-Domain Adaptive Clustering for Semi-Supervised Domain Adaptation [85.6961770631173]
In semi-supervised domain adaptation, a few labeled samples per class in the target domain guide features of the remaining target samples to aggregate around them.
We propose a novel approach called Cross-domain Adaptive Clustering to address this problem.
arXiv Detail & Related papers (2021-04-19T16:07:32Z) - Instance Level Affinity-Based Transfer for Unsupervised Domain
Adaptation [74.71931918541748]
We propose an instance affinity based criterion for source to target transfer during adaptation, called ILA-DA.
We first propose a reliable and efficient method to extract similar and dissimilar samples across source and target, and utilize a multi-sample contrastive loss to drive the domain alignment process.
We verify the effectiveness of ILA-DA by observing consistent improvements in accuracy over popular domain adaptation approaches on a variety of benchmark datasets.
arXiv Detail & Related papers (2021-04-03T01:33:14Z) - Towards Uncovering the Intrinsic Data Structures for Unsupervised Domain
Adaptation using Structurally Regularized Deep Clustering [119.88565565454378]
Unsupervised domain adaptation (UDA) is to learn classification models that make predictions for unlabeled data on a target domain.
We propose a hybrid model of Structurally Regularized Deep Clustering, which integrates the regularized discriminative clustering of target data with a generative one.
Our proposed H-SRDC outperforms all the existing methods under both the inductive and transductive settings.
arXiv Detail & Related papers (2020-12-08T08:52:00Z) - Exploring Category-Agnostic Clusters for Open-Set Domain Adaptation [138.29273453811945]
We present Self-Ensembling with Category-agnostic Clusters (SE-CC) -- a novel architecture that steers domain adaptation with category-agnostic clusters in target domain.
clustering is performed over all the unlabeled target samples to obtain the category-agnostic clusters, which reveal the underlying data space structure peculiar to target domain.
arXiv Detail & Related papers (2020-06-11T16:19:02Z) - Domain Adaptation by Class Centroid Matching and Local Manifold
Self-Learning [8.316259570013813]
We propose a novel domain adaptation approach, which can thoroughly explore the data distribution structure of target domain.
We regard the samples within the same cluster in target domain as a whole rather than individuals and assigns pseudo-labels to the target cluster by class centroid matching.
An efficient iterative optimization algorithm is designed to solve the objective function of our proposal with theoretical convergence guarantee.
arXiv Detail & Related papers (2020-03-20T16:59:27Z) - Unsupervised Domain Adaptation via Structurally Regularized Deep
Clustering [35.008158504090176]
Unsupervised domain adaptation (UDA) is to make predictions for unlabeled data on a target domain, given labeled data on a source domain whose distribution shifts from the target one.
We propose to directly uncover the intrinsic target discrimination via discriminative clustering of target data.
We term our proposed method as Structurally Regularized Deep Clustering (SRDC), where we also enhance target discrimination with clustering of intermediate network features.
arXiv Detail & Related papers (2020-03-19T07:26:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.