Evaluating Parameter-Efficient Transfer Learning Approaches on SURE
Benchmark for Speech Understanding
- URL: http://arxiv.org/abs/2303.03267v1
- Date: Thu, 2 Mar 2023 08:57:33 GMT
- Title: Evaluating Parameter-Efficient Transfer Learning Approaches on SURE
Benchmark for Speech Understanding
- Authors: Yingting Li, Ambuj Mehrish, Shuai Zhao, Rishabh Bhardwaj, Amir Zadeh,
Navonil Majumder, Rada Mihalcea, Soujanya Poria
- Abstract summary: Fine-tuning is widely used as the default algorithm for transfer learning from pre-trained models.
We introduce the Speech UndeRstanding Evaluation (SURE) benchmark for parameter-efficient learning for various speech-processing tasks.
- Score: 40.27182770995891
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Fine-tuning is widely used as the default algorithm for transfer learning
from pre-trained models. Parameter inefficiency can however arise when, during
transfer learning, all the parameters of a large pre-trained model need to be
updated for individual downstream tasks. As the number of parameters grows,
fine-tuning is prone to overfitting and catastrophic forgetting. In addition,
full fine-tuning can become prohibitively expensive when the model is used for
many tasks. To mitigate this issue, parameter-efficient transfer learning
algorithms, such as adapters and prefix tuning, have been proposed as a way to
introduce a few trainable parameters that can be plugged into large pre-trained
language models such as BERT, and HuBERT. In this paper, we introduce the
Speech UndeRstanding Evaluation (SURE) benchmark for parameter-efficient
learning for various speech-processing tasks. Additionally, we introduce a new
adapter, ConvAdapter, based on 1D convolution. We show that ConvAdapter
outperforms the standard adapters while showing comparable performance against
prefix tuning and LoRA with only 0.94% of trainable parameters on some of the
task in SURE. We further explore the effectiveness of parameter efficient
transfer learning for speech synthesis task such as Text-to-Speech (TTS).
Related papers
- Parameter-Efficient Fine-Tuning With Adapters [5.948206235442328]
This research introduces a novel adaptation method utilizing the UniPELT framework as a base.
Our method employs adapters, which enable efficient transfer of pretrained models to new tasks with minimal retraining of the base model parameters.
arXiv Detail & Related papers (2024-05-09T01:40:38Z) - Dynamic Adapter Meets Prompt Tuning: Parameter-Efficient Transfer Learning for Point Cloud Analysis [51.14136878142034]
Point cloud analysis has achieved outstanding performance by transferring point cloud pre-trained models.
Existing methods for model adaptation usually update all model parameters, which is inefficient as it relies on high computational costs.
In this paper, we aim to study parameter-efficient transfer learning for point cloud analysis with an ideal trade-off between task performance and parameter efficiency.
arXiv Detail & Related papers (2024-03-03T08:25:04Z) - Parameter-Efficient Fine-Tuning without Introducing New Latency [7.631596468553607]
We introduce a novel adapter technique that directly applies the adapter to pre-trained parameters instead of the hidden representation.
Our proposed method attains a new state-of-the-art outcome in terms of both performance and storage efficiency, storing only 0.03% parameters of full fine-tuning.
arXiv Detail & Related papers (2023-05-26T08:44:42Z) - Sensitivity-Aware Visual Parameter-Efficient Fine-Tuning [91.5113227694443]
We propose a novel visual.
sensuous-aware fine-Tuning (SPT) scheme.
SPT allocates trainable parameters to task-specific important positions.
Experiments on a wide range of downstream recognition tasks show that our SPT is complementary to the existing PEFT methods.
arXiv Detail & Related papers (2023-03-15T12:34:24Z) - Exploring Efficient-tuning Methods in Self-supervised Speech Models [53.633222197712875]
Self-supervised learning can learn powerful representations for different speech tasks.
In downstream tasks, the parameters of SSL models are frozen, and only the adapters are trained.
We show that the performance parity can be achieved with over 90% parameter reduction.
arXiv Detail & Related papers (2022-10-10T11:08:12Z) - Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than
In-Context Learning [81.3514358542452]
Few-shot in-context learning (ICL) incurs substantial computational, memory, and storage costs because it involves processing all of the training examples every time a prediction is made.
parameter-efficient fine-tuning offers an alternative paradigm where a small set of parameters are trained to enable a model to perform the new task.
In this paper, we rigorously compare few-shot ICL and parameter-efficient fine-tuning and demonstrate that the latter offers better accuracy as well as dramatically lower computational costs.
arXiv Detail & Related papers (2022-05-11T17:10:41Z) - AdapterBias: Parameter-efficient Token-dependent Representation Shift
for Adapters in NLP Tasks [55.705355299065474]
Transformer-based pre-trained models with millions of parameters require large storage.
Recent approaches tackle this shortcoming by training adapters, but these approaches still require a relatively large number of parameters.
In this study, AdapterBias, a surprisingly simple yet effective adapter architecture, is proposed.
arXiv Detail & Related papers (2022-04-30T16:49:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.