Quantifying Adaptability in Pre-trained Language Models with 500 Tasks
- URL: http://arxiv.org/abs/2112.03204v1
- Date: Mon, 6 Dec 2021 18:00:25 GMT
- Title: Quantifying Adaptability in Pre-trained Language Models with 500 Tasks
- Authors: Belinda Z. Li, Jane Yu, Madian Khabsa, Luke Zettlemoyer, Alon Halevy,
Jacob Andreas
- Abstract summary: We present a large-scale empirical study of the features and limits of LM adaptability using a new benchmark, TaskBench500.
We evaluate three facets of adaptability, finding that adaptation procedures differ dramatically in their ability to memorize small datasets.
Our experiments show that adaptability to new tasks, like generalization to new examples, can be systematically described and understood.
- Score: 60.0364822929442
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When a neural language model (LM) is adapted to perform a new task, what
aspects of the task predict the eventual performance of the model? In NLP,
systematic features of LM generalization to individual examples are well
characterized, but systematic aspects of LM adaptability to new tasks are not
nearly as well understood. We present a large-scale empirical study of the
features and limits of LM adaptability using a new benchmark, TaskBench500,
built from 500 procedurally generated sequence modeling tasks. These tasks
combine core aspects of language processing, including lexical semantics,
sequence processing, memorization, logical reasoning, and world knowledge.
Using TaskBench500, we evaluate three facets of adaptability, finding that: (1)
adaptation procedures differ dramatically in their ability to memorize small
datasets; (2) within a subset of task types, adaptation procedures exhibit
compositional adaptability to complex tasks; and (3) failure to match training
label distributions is explained by mismatches in the intrinsic difficulty of
predicting individual labels. Our experiments show that adaptability to new
tasks, like generalization to new examples, can be systematically described and
understood, and we conclude with a discussion of additional aspects of
adaptability that could be studied using the new benchmark.
Related papers
- Limits of Transformer Language Models on Learning to Compose Algorithms [77.2443883991608]
We evaluate training LLaMA models and prompting GPT-4 and Gemini on four tasks demanding to learn a composition of several discrete sub-tasks.
Our results indicate that compositional learning in state-of-the-art Transformer language models is highly sample inefficient.
arXiv Detail & Related papers (2024-02-08T16:23:29Z) - Generalization to New Sequential Decision Making Tasks with In-Context
Learning [23.36106067650874]
Training autonomous agents that can learn new tasks from only a handful of demonstrations is a long-standing problem in machine learning.
In this paper, we show that naively applying transformers to sequential decision making problems does not enable in-context learning of new tasks.
We investigate different design choices and find that larger model and dataset sizes, as well as more task diversity, environmentity, and trajectory burstiness, all result in better in-context learning of new out-of-distribution tasks.
arXiv Detail & Related papers (2023-12-06T15:19:28Z) - Towards Task Sampler Learning for Meta-Learning [37.02030832662183]
Meta-learning aims to learn general knowledge with diverse training tasks conducted from limited data, and then transfer it to new tasks.
It is commonly believed that increasing task diversity will enhance the generalization ability of meta-learning models.
This paper challenges this view through empirical and theoretical analysis.
arXiv Detail & Related papers (2023-07-18T01:53:18Z) - On Conditional and Compositional Language Model Differentiable Prompting [75.76546041094436]
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) to perform well on downstream tasks.
We propose a new model, Prompt Production System (PRopS), which learns to transform task instructions or input metadata, into continuous prompts.
arXiv Detail & Related papers (2023-07-04T02:47:42Z) - Improving Meta-learning for Low-resource Text Classification and
Generation via Memory Imitation [87.98063273826702]
We propose a memory imitation meta-learning (MemIML) method that enhances the model's reliance on support sets for task adaptation.
A theoretical analysis is provided to prove the effectiveness of our method.
arXiv Detail & Related papers (2022-03-22T12:41:55Z) - SUPERB-SG: Enhanced Speech processing Universal PERformance Benchmark
for Semantic and Generative Capabilities [76.97949110580703]
We introduce SUPERB-SG, a new benchmark to evaluate pre-trained models across various speech tasks.
We use a lightweight methodology to test the robustness of representations learned by pre-trained models under shifts in data domain.
We also show that the task diversity of SUPERB-SG coupled with limited task supervision is an effective recipe for evaluating the generalizability of model representation.
arXiv Detail & Related papers (2022-03-14T04:26:40Z) - Variable-Shot Adaptation for Online Meta-Learning [123.47725004094472]
We study the problem of learning new tasks from a small, fixed number of examples, by meta-learning across static data from a set of previous tasks.
We find that meta-learning solves the full task set with fewer overall labels and greater cumulative performance, compared to standard supervised methods.
These results suggest that meta-learning is an important ingredient for building learning systems that continuously learn and improve over a sequence of problems.
arXiv Detail & Related papers (2020-12-14T18:05:24Z) - Learning from Task Descriptions [24.588252048132862]
We introduce a framework for developing NLP systems that solve new tasks after reading their descriptions.
We instantiate this framework with a new English language dataset, ZEST, structured for task-oriented evaluation.
We find that the state-of-the-art T5 model achieves a score of 12% on ZEST, leaving a significant challenge for NLP researchers.
arXiv Detail & Related papers (2020-11-16T17:25:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.