Identifying Suitable Tasks for Inductive Transfer Through the Analysis
of Feature Attributions
- URL: http://arxiv.org/abs/2202.01096v1
- Date: Wed, 2 Feb 2022 15:51:07 GMT
- Title: Identifying Suitable Tasks for Inductive Transfer Through the Analysis
of Feature Attributions
- Authors: Alexander J. Hepburn, Richard McCreadie
- Abstract summary: We use explainability techniques to predict whether task pairs will be complementary, through comparison of neural network activation between single-task models.
Our results show that, through this approach, it is possible to reduce training time by up to 83.5% at a cost of only 0.034 reduction in positive-class F1 on the TREC-IS 2020-A dataset.
- Score: 78.55044112903148
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Transfer learning approaches have shown to significantly improve performance
on downstream tasks. However, it is common for prior works to only report where
transfer learning was beneficial, ignoring the significant trial-and-error
required to find effective settings for transfer. Indeed, not all task
combinations lead to performance benefits, and brute-force searching rapidly
becomes computationally infeasible. Hence the question arises, can we predict
whether transfer between two tasks will be beneficial without actually
performing the experiment? In this paper, we leverage explainability techniques
to effectively predict whether task pairs will be complementary, through
comparison of neural network activation between single-task models. In this
way, we can avoid grid-searches over all task and hyperparameter combinations,
dramatically reducing the time needed to find effective task pairs. Our results
show that, through this approach, it is possible to reduce training time by up
to 83.5% at a cost of only 0.034 reduction in positive-class F1 on the TREC-IS
2020-A dataset.
Related papers
- Exploring the Effectiveness and Consistency of Task Selection in Intermediate-Task Transfer Learning [21.652389166495407]
We show that the transfer performance exhibits severe variance across different source tasks and training seeds.
Compared to embedding-free methods and text embeddings, task embeddings constructed from fine-tuned weights can better estimate task transferability.
We introduce a novel method that measures pairwise token similarity using maximum inner product search, leading to the highest performance in task prediction.
arXiv Detail & Related papers (2024-07-23T07:31:43Z) - Divergence-Based Domain Transferability for Zero-Shot Classification [78.55044112903148]
Transferring learned patterns from pretrained neural language models has been shown to significantly improve effectiveness across a variety of language-based tasks.
Further tuning on intermediate tasks has been demonstrated to provide additional performance benefits, provided the intermediate task is sufficiently related to the target task.
However, how to identify related tasks is an open problem, and brute-force searching effective task combinations is prohibitively expensive.
arXiv Detail & Related papers (2023-02-11T16:04:38Z) - ForkMerge: Mitigating Negative Transfer in Auxiliary-Task Learning [59.08197876733052]
Auxiliary-Task Learning (ATL) aims to improve the performance of the target task by leveraging the knowledge obtained from related tasks.
Sometimes, learning multiple tasks simultaneously results in lower accuracy than learning only the target task, known as negative transfer.
ForkMerge is a novel approach that periodically forks the model into multiple branches, automatically searches the varying task weights.
arXiv Detail & Related papers (2023-01-30T02:27:02Z) - Transferability-Guided Cross-Domain Cross-Task Transfer Learning [21.812715282796255]
We propose two novel transferability metrics F-OTCE and JC-OTCE.
F-OTCE estimates transferability by first solving an Optimal Transport problem between source and target distributions.
JC-OTCE improves the transferability of F-OTCE by including label distances in the OT problem.
arXiv Detail & Related papers (2022-07-12T13:06:16Z) - Efficiently Identifying Task Groupings for Multi-Task Learning [55.80489920205404]
Multi-task learning can leverage information learned by one task to benefit the training of other tasks.
We suggest an approach to select which tasks should train together in multi-task learning models.
Our method determines task groupings in a single training run by co-training all tasks together and quantifying the effect to which one task's gradient would affect another task's loss.
arXiv Detail & Related papers (2021-09-10T02:01:43Z) - On-edge Multi-task Transfer Learning: Model and Practice with
Data-driven Task Allocation [20.20889051697198]
We show that task allocation with task importance for Multi-task Transfer Learning (MTL) is a variant of the NP-complete Knapsack problem.
We propose a Data-driven Cooperative Task Allocation (DCTA) approach to solve TATIM with high computational efficiency.
Our DCTA reduces 3.24 times of processing time, and saves 48.4% energy consumption compared with the state-of-the-art when solving TATIM.
arXiv Detail & Related papers (2021-07-06T08:24:25Z) - Frustratingly Easy Transferability Estimation [64.42879325144439]
We propose a simple, efficient, and effective transferability measure named TransRate.
TransRate measures the transferability as the mutual information between the features of target examples extracted by a pre-trained model and labels of them.
Despite its extraordinary simplicity in 10 lines of codes, TransRate performs remarkably well in extensive evaluations on 22 pre-trained models and 16 downstream tasks.
arXiv Detail & Related papers (2021-06-17T10:27:52Z) - Exploring and Predicting Transferability across NLP Tasks [115.6278033699853]
We study the transferability between 33 NLP tasks across three broad classes of problems.
Our results show that transfer learning is more beneficial than previously thought.
We also develop task embeddings that can be used to predict the most transferable source tasks for a given target task.
arXiv Detail & Related papers (2020-05-02T09:39:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.