OTCE: A Transferability Metric for Cross-Domain Cross-Task
Representations
- URL: http://arxiv.org/abs/2103.13843v1
- Date: Thu, 25 Mar 2021 13:51:33 GMT
- Title: OTCE: A Transferability Metric for Cross-Domain Cross-Task
Representations
- Authors: Yang Tan, Yang Li, Shao-Lun Huang
- Abstract summary: We propose a transferability metric called Optimal Transport based Conditional Entropy (OTCE)
OTCE characterizes transferability as a combination of domain difference and task difference, and explicitly evaluates them from data in a unified framework.
Experiments on the largest cross-domain dataset DomainNet and Office31 demonstrate that OTCE shows an average of 21% gain in the correlation with the ground truth transfer accuracy.
- Score: 6.730043708859326
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Transfer learning across heterogeneous data distributions (a.k.a. domains)
and distinct tasks is a more general and challenging problem than conventional
transfer learning, where either domains or tasks are assumed to be the same.
While neural network based feature transfer is widely used in transfer learning
applications, finding the optimal transfer strategy still requires
time-consuming experiments and domain knowledge. We propose a transferability
metric called Optimal Transport based Conditional Entropy (OTCE), to
analytically predict the transfer performance for supervised classification
tasks in such cross-domain and cross-task feature transfer settings. Our OTCE
score characterizes transferability as a combination of domain difference and
task difference, and explicitly evaluates them from data in a unified
framework. Specifically, we use optimal transport to estimate domain difference
and the optimal coupling between source and target distributions, which is then
used to derive the conditional entropy of the target task (task difference).
Experiments on the largest cross-domain dataset DomainNet and Office31
demonstrate that OTCE shows an average of 21% gain in the correlation with the
ground truth transfer accuracy compared to state-of-the-art methods. We also
investigate two applications of the OTCE score including source model selection
and multi-source feature fusion.
Related papers
- Transferability Metrics for Object Detection [0.0]
Transfer learning aims to make the most of existing pre-trained models to achieve better performance on a new task in limited data scenarios.
We extend transferability metrics to object detection using ROI-Align and TLogME.
We show that TLogME provides a robust correlation with transfer performance and outperforms other transferability metrics on local and global level features.
arXiv Detail & Related papers (2023-06-27T08:49:31Z) - Towards Estimating Transferability using Hard Subsets [25.86053764521497]
We propose HASTE, a new strategy to estimate the transferability of a source model to a particular target task using only a harder subset of target data.
We show that HASTE can be used with any existing transferability metric to improve their reliability.
Our experimental results across multiple source model architectures, target datasets, and transfer learning tasks show that HASTE modified metrics are consistently better or on par with the state of the art transferability metrics.
arXiv Detail & Related papers (2023-01-17T14:50:18Z) - Transferability-Guided Cross-Domain Cross-Task Transfer Learning [21.812715282796255]
We propose two novel transferability metrics F-OTCE and JC-OTCE.
F-OTCE estimates transferability by first solving an Optimal Transport problem between source and target distributions.
JC-OTCE improves the transferability of F-OTCE by including label distances in the OT problem.
arXiv Detail & Related papers (2022-07-12T13:06:16Z) - Transferability Estimation for Semantic Segmentation Task [20.07223947190349]
We extend the recent transferability metric OTCE score to the semantic segmentation task.
The challenge in applying the OTCE score is the high dimensional segmentation output, which is difficult to find the optimal coupling between so many pixels under an acceptable cost.
Experimental evaluation on Cityscapes, BDD100K and GTA5 datasets demonstrates that the OTCE score highly correlates with the transfer performance.
arXiv Detail & Related papers (2021-09-30T16:21:17Z) - Coarse to Fine: Domain Adaptive Crowd Counting via Adversarial Scoring
Network [58.05473757538834]
This paper proposes a novel adversarial scoring network (ASNet) to bridge the gap across domains from coarse to fine granularity.
Three sets of migration experiments show that the proposed methods achieve state-of-the-art counting performance.
arXiv Detail & Related papers (2021-07-27T14:47:24Z) - Practical Transferability Estimation for Image Classification Tasks [20.07223947190349]
A major challenge is how to make transfereability estimation robust under the cross-domain cross-task settings.
The recently proposed OTCE score solves this problem by considering both domain and task differences.
We propose a practical transferability metric called JC-NCE score that dramatically improves the robustness of the task difference estimation.
arXiv Detail & Related papers (2021-06-19T11:59:11Z) - Frustratingly Easy Transferability Estimation [64.42879325144439]
We propose a simple, efficient, and effective transferability measure named TransRate.
TransRate measures the transferability as the mutual information between the features of target examples extracted by a pre-trained model and labels of them.
Despite its extraordinary simplicity in 10 lines of codes, TransRate performs remarkably well in extensive evaluations on 22 pre-trained models and 16 downstream tasks.
arXiv Detail & Related papers (2021-06-17T10:27:52Z) - Latent-Optimized Adversarial Neural Transfer for Sarcasm Detection [50.29565896287595]
We apply transfer learning to exploit common datasets for sarcasm detection.
We propose a generalized latent optimization strategy that allows different losses to accommodate each other.
In particular, we achieve 10.02% absolute performance gain over the previous state of the art on the iSarcasm dataset.
arXiv Detail & Related papers (2021-04-19T13:07:52Z) - Instance Level Affinity-Based Transfer for Unsupervised Domain
Adaptation [74.71931918541748]
We propose an instance affinity based criterion for source to target transfer during adaptation, called ILA-DA.
We first propose a reliable and efficient method to extract similar and dissimilar samples across source and target, and utilize a multi-sample contrastive loss to drive the domain alignment process.
We verify the effectiveness of ILA-DA by observing consistent improvements in accuracy over popular domain adaptation approaches on a variety of benchmark datasets.
arXiv Detail & Related papers (2021-04-03T01:33:14Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z) - Domain Conditioned Adaptation Network [90.63261870610211]
We propose a Domain Conditioned Adaptation Network (DCAN) to excite distinct convolutional channels with a domain conditioned channel attention mechanism.
This is the first work to explore the domain-wise convolutional channel activation for deep DA networks.
arXiv Detail & Related papers (2020-05-14T04:23:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.