Parameter Efficient Finetuning for Speech Emotion Recognition and Domain
Adaptation
- URL: http://arxiv.org/abs/2402.11747v1
- Date: Mon, 19 Feb 2024 00:21:07 GMT
- Title: Parameter Efficient Finetuning for Speech Emotion Recognition and Domain
Adaptation
- Authors: Nineli Lashkarashvili, Wen Wu, Guangzhi Sun, Philip C. Woodland
- Abstract summary: This paper investigates parameter-efficient finetuning (PEFT) for speech emotion recognition (SER)
Various PEFT adaptors are systematically studied for both classification of discrete emotion categories and prediction of dimensional emotional attributes.
A two-stage adaptation strategy is proposed to adapt models trained on acted emotion data, which is more readily available, to make the model more adept at capturing natural emotional expressions.
- Score: 13.774287532165019
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Foundation models have shown superior performance for speech emotion
recognition (SER). However, given the limited data in emotion corpora,
finetuning all parameters of large pre-trained models for SER can be both
resource-intensive and susceptible to overfitting. This paper investigates
parameter-efficient finetuning (PEFT) for SER. Various PEFT adaptors are
systematically studied for both classification of discrete emotion categories
and prediction of dimensional emotional attributes. The results demonstrate
that the combination of PEFT methods surpasses full finetuning with a
significant reduction in the number of trainable parameters. Furthermore, a
two-stage adaptation strategy is proposed to adapt models trained on acted
emotion data, which is more readily available, to make the model more adept at
capturing natural emotional expressions. Both intra- and cross-corpus
experiments validate the efficacy of the proposed approach in enhancing the
performance on both the source and target domains.
Related papers
- EEG-Based Mental Imagery Task Adaptation via Ensemble of Weight-Decomposed Low-Rank Adapters [8.238310169506944]
We propose a novel ensemble of weight-decomposed low-rank adaptation methods, EDoRA, for parameter-efficient mental imagery task adaptation.
The proposed method has performed better than full fine-tune and state-of-the-art PEFT methods for mental imagery classification.
arXiv Detail & Related papers (2024-12-08T14:57:00Z) - Preserving Pre-trained Representation Space: On Effectiveness of Prefix-tuning for Large Multi-modal Models [24.62337386603331]
Large Multi-modal Models (LMMs) are revolutionizing the way machines interact with the world.
To adapt LMMs for downstream tasks, parameter-efficient fine-tuning (PEFT) has gained popularity.
This paper focuses on the strengths and weaknesses of each tuning strategy, shifting the focus from the efficiency typically associated with these approaches.
arXiv Detail & Related papers (2024-10-29T07:55:50Z) - ETHER: Efficient Finetuning of Large-Scale Models with Hyperplane Reflections [59.839926875976225]
We propose the ETHER transformation family, which performs Efficient fineTuning via HypErplane Reflections.
In particular, we introduce ETHER and its relaxation ETHER+, which match or outperform existing PEFT methods with significantly fewer parameters.
arXiv Detail & Related papers (2024-05-30T17:26:02Z) - Parameter Efficient Fine Tuning: A Comprehensive Analysis Across Applications [0.7421845364041001]
The rise of deep learning has marked significant progress in fields such as computer vision, natural language processing, and medical imaging.
Traditional fine-tuning methods, involving adjustments to all parameters, face challenges due to high computational and memory demands.
This review examines Efficient Fine-Tuning (PEFT) techniques, which selectively update parameters to balance computational efficiency with performance.
arXiv Detail & Related papers (2024-04-21T02:26:15Z) - Dynamic Tuning Towards Parameter and Inference Efficiency for ViT Adaptation [67.13876021157887]
Dynamic Tuning (DyT) is a novel approach to improve both parameter and inference efficiency for ViT adaptation.
DyT achieves superior performance compared to existing PEFT methods while evoking only 71% of their FLOPs on the VTAB-1K benchmark.
arXiv Detail & Related papers (2024-03-18T14:05:52Z) - Advancing Parameter Efficiency in Fine-tuning via Representation Editing [41.81020951061438]
We propose a novel fine-tuning approach for neural models, named Representation EDiting (RED)
RED modifies the representations generated at some layers through the application of scaling and biasing operations.
Remarkably, RED achieves results comparable or superior to both full parameter fine-tuning and other PEFT methods.
arXiv Detail & Related papers (2024-02-23T08:21:02Z) - Bayesian Parameter-Efficient Fine-Tuning for Overcoming Catastrophic Forgetting [10.559392015748989]
We show that catastrophic forgetting can be overcome by our methods without degrading the fine-tuning performance.
Our results demonstrate that using the Kronecker-factored approximation produces a better preservation of the pre-training knowledge than the diagonal ones.
arXiv Detail & Related papers (2024-02-19T15:26:19Z) - E^2VPT: An Effective and Efficient Approach for Visual Prompt Tuning [55.50908600818483]
Fine-tuning large-scale pretrained vision models for new tasks has become increasingly parameter-intensive.
We propose an Effective and Efficient Visual Prompt Tuning (E2VPT) approach for large-scale transformer-based model adaptation.
Our approach outperforms several state-of-the-art baselines on two benchmarks.
arXiv Detail & Related papers (2023-07-25T19:03:21Z) - An Empirical Analysis of Parameter-Efficient Methods for Debiasing
Pre-Trained Language Models [55.14405248920852]
We conduct experiments with prefix tuning, prompt tuning, and adapter tuning on different language models and bias types to evaluate their debiasing performance.
We find that the parameter-efficient methods are effective in mitigating gender bias, where adapter tuning is consistently the most effective.
We also find that prompt tuning is more suitable for GPT-2 than BERT, and racial and religious bias is less effective when it comes to racial and religious bias.
arXiv Detail & Related papers (2023-06-06T23:56:18Z) - Towards a Unified View on Visual Parameter-Efficient Transfer Learning [96.99924127527002]
We propose a framework with a unified view called visual-PETL (V-PETL) to investigate the different aspects affecting the trade-off.
An effective scheme Swin-BAPAT derived from the proposed V-PETL framework achieves significantly better performance than the state-of-the-art AdaptFormer-Swin.
arXiv Detail & Related papers (2022-10-03T09:54:39Z) - Delta Tuning: A Comprehensive Study of Parameter Efficient Methods for
Pre-trained Language Models [90.24999406296867]
In contrast with the standard fine-tuning, delta tuning only fine-tunes a small portion of the model parameters while keeping the rest untouched.
Recent studies have demonstrated that a series of delta tuning methods with distinct tuned parameter selection could achieve performance on a par with full- parameter fine-tuning.
arXiv Detail & Related papers (2022-03-14T07:56:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.