Transferability Metrics for Selecting Source Model Ensembles
- URL: http://arxiv.org/abs/2111.13011v1
- Date: Thu, 25 Nov 2021 10:43:29 GMT
- Title: Transferability Metrics for Selecting Source Model Ensembles
- Authors: Andrea Agostinelli, Jasper Uijlings, Thomas Mensink, Vittorio Ferrari
- Abstract summary: ensemble selection is difficult because fine-tuning all possible ensembles is computationally prohibitive.
We propose several new transferability metrics designed for this task and evaluate them in a challenging and realistic transfer learning setup.
Averaged over 17 target datasets, we outperform these baselines by 6.4% and 2.5% relative mean IoU, respectively.
- Score: 43.980600479738435
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We address the problem of ensemble selection in transfer learning: Given a
large pool of source models we want to select an ensemble of models which,
after fine-tuning on the target training set, yields the best performance on
the target test set. Since fine-tuning all possible ensembles is
computationally prohibitive, we aim at predicting performance on the target
dataset using a computationally efficient transferability metric. We propose
several new transferability metrics designed for this task and evaluate them in
a challenging and realistic transfer learning setup for semantic segmentation:
we create a large and diverse pool of source models by considering 17 source
datasets covering a wide variety of image domain, two different architectures,
and two pre-training schemes. Given this pool, we then automatically select a
subset to form an ensemble performing well on a given target dataset. We
compare the ensemble selected by our method to two baselines which select a
single source model, either (1) from the same pool as our method; or (2) from a
pool containing large source models, each with similar capacity as an ensemble.
Averaged over 17 target datasets, we outperform these baselines by 6.0% and
2.5% relative mean IoU, respectively.
Related papers
- Adapt-$\infty$: Scalable Lifelong Multimodal Instruction Tuning via Dynamic Data Selection [89.42023974249122]
Adapt-$infty$ is a new multi-way and adaptive data selection approach for Lifelong Instruction Tuning.
We construct pseudo-skill clusters by grouping gradient-based sample vectors.
We select the best-performing data selector for each skill cluster from a pool of selector experts.
arXiv Detail & Related papers (2024-10-14T15:48:09Z) - DsDm: Model-Aware Dataset Selection with Datamodels [81.01744199870043]
Standard practice is to filter for examples that match human notions of data quality.
We find that selecting according to similarity with "high quality" data sources may not increase (and can even hurt) performance compared to randomly selecting data.
Our framework avoids handpicked notions of data quality, and instead models explicitly how the learning process uses train datapoints to predict on the target tasks.
arXiv Detail & Related papers (2024-01-23T17:22:00Z) - Large Language Model Routing with Benchmark Datasets [40.42044096089315]
No single model typically achieves the best accuracy in all tasks and use cases.
We propose a new formulation for the problem, in which benchmark datasets are repurposed to learn a "router" model for this selection.
We show that this problem can be reduced to a collection of binary classification tasks.
arXiv Detail & Related papers (2023-09-27T17:08:40Z) - Building a Winning Team: Selecting Source Model Ensembles using a
Submodular Transferability Estimation Approach [20.86345962679122]
Estimating the transferability of publicly available pretrained models to a target task has assumed an important place for transfer learning tasks.
We propose a novel Optimal tranSport-based suBmOdular tRaNsferability metric (OSBORN) to estimate the transferability of an ensemble of models to a downstream task.
arXiv Detail & Related papers (2023-09-05T17:57:31Z) - Universal Semi-supervised Model Adaptation via Collaborative Consistency
Training [92.52892510093037]
We introduce a realistic and challenging domain adaptation problem called Universal Semi-supervised Model Adaptation (USMA)
We propose a collaborative consistency training framework that regularizes the prediction consistency between two models.
Experimental results demonstrate the effectiveness of our method on several benchmark datasets.
arXiv Detail & Related papers (2023-07-07T08:19:40Z) - Deep Model Reassembly [60.6531819328247]
We explore a novel knowledge-transfer task, termed as Deep Model Reassembly (DeRy)
The goal of DeRy is to first dissect each model into distinctive building blocks, and then selectively reassemble the derived blocks to produce customized networks.
We demonstrate that on ImageNet, the best reassemble model achieves 78.6% top-1 accuracy without fine-tuning.
arXiv Detail & Related papers (2022-10-24T10:16:13Z) - Single-dataset Experts for Multi-dataset Question Answering [6.092171111087768]
We train a network on multiple datasets to generalize and transfer better to new datasets.
Our approach is to model multi-dataset question answering with a collection of single-dataset experts.
Simple methods based on parameter-averaging lead to better zero-shot generalization and few-shot transfer performance.
arXiv Detail & Related papers (2021-09-28T17:08:22Z) - On Generalization in Coreference Resolution [66.05112218880907]
We consolidate a set of 8 coreference resolution datasets targeting different domains to evaluate the off-the-shelf performance of models.
We then mix three datasets for training; even though their domain, annotation guidelines, and metadata differ, we propose a method for jointly training a single model.
We find that in a zero-shot setting, models trained on a single dataset transfer poorly while joint training yields improved overall performance.
arXiv Detail & Related papers (2021-09-20T16:33:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.