Exploiting a Zoo of Checkpoints for Unseen Tasks
- URL: http://arxiv.org/abs/2111.03628v1
- Date: Fri, 5 Nov 2021 17:27:22 GMT
- Title: Exploiting a Zoo of Checkpoints for Unseen Tasks
- Authors: Jiaji Huang, Qiang Qiu, Kenneth Church
- Abstract summary: We model the space of tasks as a Gaussian process.
We can identify representative checkpoints by a maximum mutual information criterion.
A greedy method identifies representatives that are likely to "cover" the task space.
- Score: 29.309248850221373
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: There are so many models in the literature that it is difficult for
practitioners to decide which combinations are likely to be effective for a new
task. This paper attempts to address this question by capturing relationships
among checkpoints published on the web. We model the space of tasks as a
Gaussian process. The covariance can be estimated from checkpoints and
unlabeled probing data. With the Gaussian process, we can identify
representative checkpoints by a maximum mutual information criterion. This
objective is submodular. A greedy method identifies representatives that are
likely to "cover" the task space. These representatives generalize to new tasks
with superior performance. Empirical evidence is provided for applications from
both computational linguistics as well as computer vision.
Related papers
- Data-CUBE: Data Curriculum for Instruction-based Sentence Representation
Learning [85.66907881270785]
We propose a data curriculum method, namely Data-CUBE, that arranges the orders of all the multi-task data for training.
In the task level, we aim to find the optimal task order to minimize the total cross-task interference risk.
In the instance level, we measure the difficulty of all instances per task, then divide them into the easy-to-difficult mini-batches for training.
arXiv Detail & Related papers (2024-01-07T18:12:20Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - Is Probing All You Need? Indicator Tasks as an Alternative to Probing
Embedding Spaces [19.4968960182412]
We introduce the term indicator tasks for non-trainable tasks which are used to query embedding spaces for the existence of certain properties.
We show that the application of a suitable indicator provides a more accurate picture of the information captured and removed compared to probes.
arXiv Detail & Related papers (2023-10-24T15:08:12Z) - Boosting Multitask Learning on Graphs through Higher-Order Task Affinities [17.70434437597516]
Predicting node labels on a given graph is a widely studied problem with many applications, including community detection and molecular graph prediction.
This paper considers predicting multiple node labeling functions on graphs simultaneously and revisits this problem from a multitask learning perspective.
We develop an algorithm to cluster tasks into groups based on a higher-order task affinity measure.
arXiv Detail & Related papers (2023-06-24T15:53:38Z) - Task Discovery: Finding the Tasks that Neural Networks Generalize on [1.4043229953691112]
We show that one set of images can give rise to many tasks on which neural networks generalize well.
As an example, we show that the discovered tasks can be used to automatically create adversarial train-test splits.
arXiv Detail & Related papers (2022-12-01T03:57:48Z) - MURAL: Meta-Learning Uncertainty-Aware Rewards for Outcome-Driven
Reinforcement Learning [65.52675802289775]
We show that an uncertainty aware classifier can solve challenging reinforcement learning problems.
We propose a novel method for computing the normalized maximum likelihood (NML) distribution.
We show that the resulting algorithm has a number of intriguing connections to both count-based exploration methods and prior algorithms for learning reward functions.
arXiv Detail & Related papers (2021-07-15T08:19:57Z) - Conditional Meta-Learning of Linear Representations [57.90025697492041]
Standard meta-learning for representation learning aims to find a common representation to be shared across multiple tasks.
In this work we overcome this issue by inferring a conditioning function, mapping the tasks' side information into a representation tailored to the task at hand.
We propose a meta-algorithm capable of leveraging this advantage in practice.
arXiv Detail & Related papers (2021-03-30T12:02:14Z) - Adaptive Task Sampling for Meta-Learning [79.61146834134459]
Key idea of meta-learning for few-shot classification is to mimic the few-shot situations faced at test time.
We propose an adaptive task sampling method to improve the generalization performance.
arXiv Detail & Related papers (2020-07-17T03:15:53Z) - Sequential Transfer in Reinforcement Learning with a Generative Model [48.40219742217783]
We show how to reduce the sample complexity for learning new tasks by transferring knowledge from previously-solved ones.
We derive PAC bounds on its sample complexity which clearly demonstrate the benefits of using this kind of prior knowledge.
We empirically verify our theoretical findings in simple simulated domains.
arXiv Detail & Related papers (2020-07-01T19:53:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.