Enhancing Information Maximization with Distance-Aware Contrastive
Learning for Source-Free Cross-Domain Few-Shot Learning
- URL: http://arxiv.org/abs/2403.01966v1
- Date: Mon, 4 Mar 2024 12:10:24 GMT
- Title: Enhancing Information Maximization with Distance-Aware Contrastive
Learning for Source-Free Cross-Domain Few-Shot Learning
- Authors: Huali Xu, Li Liu, Shuaifeng Zhi, Shaojing Fu, Zhuo Su, Ming-Ming
Cheng, Yongxiang Liu
- Abstract summary: Cross-Domain Few-Shot Learning methods require access to source domain data to train a model in the pre-training phase.
Due to increasing concerns about data privacy and the desire to reduce data transmission and training costs, it is necessary to develop a CDFSL solution without accessing source data.
This paper proposes an Enhanced Information Maximization with Distance-Aware Contrastive Learning method to address these challenges.
- Score: 55.715623885418815
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing Cross-Domain Few-Shot Learning (CDFSL) methods require access to
source domain data to train a model in the pre-training phase. However, due to
increasing concerns about data privacy and the desire to reduce data
transmission and training costs, it is necessary to develop a CDFSL solution
without accessing source data. For this reason, this paper explores a
Source-Free CDFSL (SF-CDFSL) problem, in which CDFSL is addressed through the
use of existing pretrained models instead of training a model with source data,
avoiding accessing source data. This paper proposes an Enhanced Information
Maximization with Distance-Aware Contrastive Learning (IM-DCL) method to
address these challenges. Firstly, we introduce the transductive mechanism for
learning the query set. Secondly, information maximization (IM) is explored to
map target samples into both individual certainty and global diversity
predictions, helping the source model better fit the target data distribution.
However, IM fails to learn the decision boundary of the target task. This
motivates us to introduce a novel approach called Distance-Aware Contrastive
Learning (DCL), in which we consider the entire feature set as both positive
and negative sets, akin to Schrodinger's concept of a dual state. Instead of a
rigid separation between positive and negative sets, we employ a weighted
distance calculation among features to establish a soft classification of the
positive and negative sets for the entire feature set. Furthermore, we address
issues related to IM by incorporating contrastive constraints between object
features and their corresponding positive and negative sets. Evaluations of the
4 datasets in the BSCD-FSL benchmark indicate that the proposed IM-DCL, without
accessing the source domain, demonstrates superiority over existing methods,
especially in the distant domain task.
Related papers
- Step-wise Distribution Alignment Guided Style Prompt Tuning for Source-free Cross-domain Few-shot Learning [53.60934432718044]
Cross-domain few-shot learning methods face challenges with large-scale pre-trained models due to inaccessible source data and training strategies.
This paper introduces Step-wise Distribution Alignment Guided Style Prompt Tuning (StepSPT)
StepSPT implicitly narrows domain gaps through prediction distribution optimization.
arXiv Detail & Related papers (2024-11-15T09:34:07Z) - Empowering Source-Free Domain Adaptation with MLLM-driven Curriculum Learning [5.599218556731767]
Source-Free Domain Adaptation (SFDA) aims to adapt a pre-trained source model to a target domain using only unlabeled target data.
Reliability-based Curriculum Learning (RCL) integrates multiple MLLMs for knowledge exploitation via pseudo-labeling in SFDA.
arXiv Detail & Related papers (2024-05-28T17:18:17Z) - Unified Source-Free Domain Adaptation [44.95240684589647]
In pursuit of transferring a source model to a target domain without access to the source training data, Source-Free Domain Adaptation (SFDA) has been extensively explored.
We propose a novel approach called Latent Causal Factors Discovery (LCFD)
In contrast to previous alternatives that emphasize learning the statistical description of reality, we formulate LCFD from a causality perspective.
arXiv Detail & Related papers (2024-03-12T12:40:08Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Learning Invariant Representation with Consistency and Diversity for
Semi-supervised Source Hypothesis Transfer [46.68586555288172]
We propose a novel task named Semi-supervised Source Hypothesis Transfer (SSHT), which performs domain adaptation based on source trained model, to generalize well in target domain with a few supervisions.
We propose Consistency and Diversity Learning (CDL), a simple but effective framework for SSHT by facilitating prediction consistency between two randomly augmented unlabeled data.
Experimental results show that our method outperforms existing SSDA methods and unsupervised model adaptation methods on DomainNet, Office-Home and Office-31 datasets.
arXiv Detail & Related papers (2021-07-07T04:14:24Z) - Source Data-absent Unsupervised Domain Adaptation through Hypothesis
Transfer and Labeling Transfer [137.36099660616975]
Unsupervised adaptation adaptation (UDA) aims to transfer knowledge from a related but different well-labeled source domain to a new unlabeled target domain.
Most existing UDA methods require access to the source data, and thus are not applicable when the data are confidential and not shareable due to privacy concerns.
This paper aims to tackle a realistic setting with only a classification model available trained over, instead of accessing to the source data.
arXiv Detail & Related papers (2020-12-14T07:28:50Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z) - Do We Really Need to Access the Source Data? Source Hypothesis Transfer
for Unsupervised Domain Adaptation [102.67010690592011]
Unsupervised adaptationUDA (UDA) aims to leverage the knowledge learned from a labeled source dataset to solve similar tasks in a new unlabeled domain.
Prior UDA methods typically require to access the source data when learning to adapt the model.
This work tackles a practical setting where only a trained source model is available and how we can effectively utilize such a model without source data to solve UDA problems.
arXiv Detail & Related papers (2020-02-20T03:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.