Black-box Source-free Domain Adaptation via Two-stage Knowledge
Distillation
- URL: http://arxiv.org/abs/2305.07881v3
- Date: Wed, 23 Aug 2023 14:53:59 GMT
- Title: Black-box Source-free Domain Adaptation via Two-stage Knowledge
Distillation
- Authors: Shuai Wang, Daoan Zhang, Zipei Yan, Shitong Shao, Rui Li
- Abstract summary: Source-free domain adaptation aims to adapt deep neural networks using only pre-trained source models and target data.
accessing the source model still has a potential concern about leaking the source data, which reveals the patient's privacy.
We study the challenging but practical problem: black-box source-free domain adaptation where only the outputs of the source model and target data are available.
- Score: 8.224874938178633
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Source-free domain adaptation aims to adapt deep neural networks using only
pre-trained source models and target data. However, accessing the source model
still has a potential concern about leaking the source data, which reveals the
patient's privacy. In this paper, we study the challenging but practical
problem: black-box source-free domain adaptation where only the outputs of the
source model and target data are available. We propose a simple but effective
two-stage knowledge distillation method. In Stage
\uppercase\expandafter{\romannumeral1}, we train the target model from scratch
with soft pseudo-labels generated by the source model in a knowledge
distillation manner. In Stage \uppercase\expandafter{\romannumeral2}, we
initialize another model as the new student model to avoid the error
accumulation caused by noisy pseudo-labels. We feed the images with weak
augmentation to the teacher model to guide the learning of the student model.
Our method is simple and flexible, and achieves surprising results on three
cross-domain segmentation tasks.
Related papers
- Cross-Domain Transfer Learning with CoRTe: Consistent and Reliable
Transfer from Black-Box to Lightweight Segmentation Model [25.3403116022412]
CoRTe is a pseudo-labelling function that extracts reliable knowledge from a black-box source model.
We benchmark CoRTe on two synthetic-to-real settings, demonstrating remarkable results when using black-box models to transfer knowledge on lightweight models for a target data distribution.
arXiv Detail & Related papers (2024-02-20T16:35:14Z) - Learn to Unlearn for Deep Neural Networks: Minimizing Unlearning
Interference with Gradient Projection [56.292071534857946]
Recent data-privacy laws have sparked interest in machine unlearning.
Challenge is to discard information about the forget'' data without altering knowledge about remaining dataset.
We adopt a projected-gradient based learning method, named as Projected-Gradient Unlearning (PGU)
We provide empirically evidence to demonstrate that our unlearning method can produce models that behave similar to models retrained from scratch across various metrics even when the training dataset is no longer accessible.
arXiv Detail & Related papers (2023-12-07T07:17:24Z) - Towards Few-Call Model Stealing via Active Self-Paced Knowledge
Distillation and Diffusion-Based Image Generation [33.60710287553274]
We propose to copy black-box classification models without having access to the original training data, the architecture, and the weights of the model.
We employ a novel active self-paced learning framework to make the most of the proxy data during distillation.
Our empirical results on two data sets confirm the superiority of our framework over two state-of-the-art methods in the few-call model extraction scenario.
arXiv Detail & Related papers (2023-09-29T19:09:27Z) - Phasic Content Fusing Diffusion Model with Directional Distribution
Consistency for Few-Shot Model Adaption [73.98706049140098]
We propose a novel phasic content fusing few-shot diffusion model with directional distribution consistency loss.
Specifically, we design a phasic training strategy with phasic content fusion to help our model learn content and style information when t is large.
Finally, we propose a cross-domain structure guidance strategy that enhances structure consistency during domain adaptation.
arXiv Detail & Related papers (2023-09-07T14:14:11Z) - Privacy Distillation: Reducing Re-identification Risk of Multimodal
Diffusion Models [11.659461421660613]
We introduce Privacy Distillation, a framework that allows a text-to-image generative model to teach another model without exposing it to identifiable data.
Our solution consists of (1) training a first diffusion model on real data (2) generating a synthetic dataset using this model and filtering it to exclude images with a re-identifiability risk (3) training a second diffusion model on the filtered synthetic data only.
arXiv Detail & Related papers (2023-06-02T07:44:00Z) - A Curriculum-style Self-training Approach for Source-Free Semantic Segmentation [91.13472029666312]
We propose a curriculum-style self-training approach for source-free domain adaptive semantic segmentation.
Our method yields state-of-the-art performance on source-free semantic segmentation tasks for both synthetic-to-real and adverse conditions.
arXiv Detail & Related papers (2021-06-22T10:21:39Z) - Distill and Fine-tune: Effective Adaptation from a Black-box Source
Model [138.12678159620248]
Unsupervised domain adaptation (UDA) aims to transfer knowledge in previous related labeled datasets (source) to a new unlabeled dataset (target)
We propose a novel two-step adaptation framework called Distill and Fine-tune (Dis-tune)
arXiv Detail & Related papers (2021-04-04T05:29:05Z) - Self-Supervised Noisy Label Learning for Source-Free Unsupervised Domain
Adaptation [87.60688582088194]
We propose a novel Self-Supervised Noisy Label Learning method.
Our method can easily achieve state-of-the-art results and surpass other methods by a very large margin.
arXiv Detail & Related papers (2021-02-23T10:51:45Z) - An Efficient Method of Training Small Models for Regression Problems
with Knowledge Distillation [1.433758865948252]
We propose a new formalism of knowledge distillation for regression problems.
First, we propose a new loss function, teacher outlier loss rejection, which rejects outliers in training samples using teacher model predictions.
By considering the multi-task network, training of the feature extraction of student models becomes more effective.
arXiv Detail & Related papers (2020-02-28T08:46:12Z) - Do We Really Need to Access the Source Data? Source Hypothesis Transfer
for Unsupervised Domain Adaptation [102.67010690592011]
Unsupervised adaptationUDA (UDA) aims to leverage the knowledge learned from a labeled source dataset to solve similar tasks in a new unlabeled domain.
Prior UDA methods typically require to access the source data when learning to adapt the model.
This work tackles a practical setting where only a trained source model is available and how we can effectively utilize such a model without source data to solve UDA problems.
arXiv Detail & Related papers (2020-02-20T03:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.