Unsupervised Finetuning
- URL: http://arxiv.org/abs/2110.09510v1
- Date: Mon, 18 Oct 2021 17:57:05 GMT
- Title: Unsupervised Finetuning
- Authors: Suichan Li and Dongdong Chen and Yinpeng Chen and Lu Yuan and Lei
Zhang and Qi Chu and Bin Liu and Nenghai Yu
- Abstract summary: We propose two strategies to combine source and target data into unsupervised finetuning.
The motivation of the former strategy is to add a small portion of source data back to occupy their pretrained representation space.
The motivation of the latter strategy is to increase the data density and help learn more compact representation.
- Score: 80.58625921631506
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies "unsupervised finetuning", the symmetrical problem of the
well-known "supervised finetuning". Given a pretrained model and small-scale
unlabeled target data, unsupervised finetuning is to adapt the representation
pretrained from the source domain to the target domain so that better transfer
performance can be obtained. This problem is more challenging than the
supervised counterpart, as the low data density in the small-scale target data
is not friendly for unsupervised learning, leading to the damage of the
pretrained representation and poor representation in the target domain. In this
paper, we find the source data is crucial when shifting the finetuning paradigm
from supervise to unsupervise, and propose two simple and effective strategies
to combine source and target data into unsupervised finetuning: "sparse source
data replaying", and "data mixing". The motivation of the former strategy is to
add a small portion of source data back to occupy their pretrained
representation space and help push the target data to reside in a smaller
compact space; and the motivation of the latter strategy is to increase the
data density and help learn more compact representation. To demonstrate the
effectiveness of our proposed ``unsupervised finetuning'' strategy, we conduct
extensive experiments on multiple different target datasets, which show better
transfer performance than the naive strategy.
Related papers
- Data Pruning in Generative Diffusion Models [2.0111637969968]
Generative models aim to estimate the underlying distribution of the data, so presumably they should benefit from larger datasets.
We show that eliminating redundant or noisy data in large datasets is beneficial particularly when done strategically.
arXiv Detail & Related papers (2024-11-19T14:13:25Z) - TCGU: Data-centric Graph Unlearning based on Transferable Condensation [36.670771080732486]
Transferable Condensation Graph Unlearning (TCGU) is a data-centric solution to zero-glance graph unlearning.
We show that TCGU can achieve superior performance in terms of model utility, unlearning efficiency, and unlearning efficacy than existing GU methods.
arXiv Detail & Related papers (2024-10-09T02:14:40Z) - The Power and Limitation of Pretraining-Finetuning for Linear Regression
under Covariate Shift [127.21287240963859]
We investigate a transfer learning approach with pretraining on the source data and finetuning based on the target data.
For a large class of linear regression instances, transfer learning with $O(N2)$ source data is as effective as supervised learning with $N$ target data.
arXiv Detail & Related papers (2022-08-03T05:59:49Z) - Unsupervised Adaptation of Semantic Segmentation Models without Source
Data [14.66682099621276]
We consider the novel problem of unsupervised domain adaptation of source models, without access to the source data for semantic segmentation.
We propose a self-training approach to extract the knowledge from the source model.
Our framework is able to achieve significant performance gains compared to directly applying the source model on the target data.
arXiv Detail & Related papers (2021-12-04T15:13:41Z) - Domain-Irrelevant Representation Learning for Unsupervised Domain
Generalization [22.980607134596077]
Domain generalization (DG) aims to help models trained on a set of source domains generalize better on unseen target domains.
While unlabeled data are far more accessible, we seek to explore how unsupervised learning can help deep models generalizes across domains.
We propose a Domain-Irrelevant Unsupervised Learning (DIUL) method to cope with the significant and misleading heterogeneity within unlabeled data.
arXiv Detail & Related papers (2021-07-13T16:20:50Z) - Delving into Data: Effectively Substitute Training for Black-box Attack [84.85798059317963]
We propose a novel perspective substitute training that focuses on designing the distribution of data used in the knowledge stealing process.
The combination of these two modules can further boost the consistency of the substitute model and target model, which greatly improves the effectiveness of adversarial attack.
arXiv Detail & Related papers (2021-04-26T07:26:29Z) - Distill and Fine-tune: Effective Adaptation from a Black-box Source
Model [138.12678159620248]
Unsupervised domain adaptation (UDA) aims to transfer knowledge in previous related labeled datasets (source) to a new unlabeled dataset (target)
We propose a novel two-step adaptation framework called Distill and Fine-tune (Dis-tune)
arXiv Detail & Related papers (2021-04-04T05:29:05Z) - Regressive Domain Adaptation for Unsupervised Keypoint Detection [67.2950306888855]
Domain adaptation (DA) aims at transferring knowledge from a labeled source domain to an unlabeled target domain.
We present a method of regressive domain adaptation (RegDA) for unsupervised keypoint detection.
Our method brings large improvement by 8% to 11% in terms of PCK on different datasets.
arXiv Detail & Related papers (2021-03-10T16:45:22Z) - Negative Data Augmentation [127.28042046152954]
We show that negative data augmentation samples provide information on the support of the data distribution.
We introduce a new GAN training objective where we use NDA as an additional source of synthetic data for the discriminator.
Empirically, models trained with our method achieve improved conditional/unconditional image generation along with improved anomaly detection capabilities.
arXiv Detail & Related papers (2021-02-09T20:28:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.