Communication-Efficient Personalized Federated Learning for
Speech-to-Text Tasks
- URL: http://arxiv.org/abs/2401.10070v1
- Date: Thu, 18 Jan 2024 15:39:38 GMT
- Title: Communication-Efficient Personalized Federated Learning for
Speech-to-Text Tasks
- Authors: Yichao Du, Zhirui Zhang, Linan Yue, Xu Huang, Yuqing Zhang, Tong Xu,
Linli Xu and Enhong Chen
- Abstract summary: To protect privacy and meet legal regulations, federated learning (FL) has gained significant attention for training speech-to-text (S2T) systems.
The commonly used FL approach (i.e., textscFedAvg) in S2T tasks typically suffers from extensive communication overhead.
We propose a personalized federated S2T framework that introduces textscFedLoRA, a lightweight LoRA module for client-side tuning and interaction with the server, and textscFedMem, a global model equipped with a $k$-near
- Score: 66.78640306687227
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To protect privacy and meet legal regulations, federated learning (FL) has
gained significant attention for training speech-to-text (S2T) systems,
including automatic speech recognition (ASR) and speech translation (ST).
However, the commonly used FL approach (i.e., \textsc{FedAvg}) in S2T tasks
typically suffers from extensive communication overhead due to multi-round
interactions based on the whole model and performance degradation caused by
data heterogeneity among clients.To address these issues, we propose a
personalized federated S2T framework that introduces \textsc{FedLoRA}, a
lightweight LoRA module for client-side tuning and interaction with the server
to minimize communication overhead, and \textsc{FedMem}, a global model
equipped with a $k$-nearest-neighbor ($k$NN) classifier that captures
client-specific distributional shifts to achieve personalization and overcome
data heterogeneity. Extensive experiments based on Conformer and Whisper
backbone models on CoVoST and GigaSpeech benchmarks show that our approach
significantly reduces the communication overhead on all S2T tasks and
effectively personalizes the global model to overcome data heterogeneity.
Related papers
- Personalized federated learning based on feature fusion [2.943623084019036]
Federated learning enables distributed clients to collaborate on training while storing their data locally to protect client privacy.
We propose a personalized federated learning approach called pFedPM.
In our process, we replace traditional gradient uploading with feature uploading, which helps reduce communication costs and allows for heterogeneous client models.
arXiv Detail & Related papers (2024-06-24T12:16:51Z) - Personalized Wireless Federated Learning for Large Language Models [75.22457544349668]
Large Language Models (LLMs) have revolutionized natural language processing tasks.
Their deployment in wireless networks still face challenges, i.e., a lack of privacy and security protection mechanisms.
We introduce two personalized wireless federated fine-tuning methods with low communication overhead.
arXiv Detail & Related papers (2024-04-20T02:30:21Z) - Towards Instance-adaptive Inference for Federated Learning [80.38701896056828]
Federated learning (FL) is a distributed learning paradigm that enables multiple clients to learn a powerful global model by aggregating local training.
In this paper, we present a novel FL algorithm, i.e., FedIns, to handle intra-client data heterogeneity by enabling instance-adaptive inference in the FL framework.
Our experiments show that our FedIns outperforms state-of-the-art FL algorithms, e.g., a 6.64% improvement against the top-performing method with less than 15% communication cost on Tiny-ImageNet.
arXiv Detail & Related papers (2023-08-11T09:58:47Z) - Collaborative Chinese Text Recognition with Personalized Federated
Learning [61.34060587461462]
In Chinese text recognition, it is often necessary for one organization to collect a large amount of data from similar organizations.
Due to the natural presence of private information in text data, such as addresses and phone numbers, different organizations are unwilling to share private data.
We introduce personalized federated learning (pFL) into the Chinese text recognition task and propose the pFedCR algorithm.
arXiv Detail & Related papers (2023-05-09T16:51:00Z) - DisPFL: Towards Communication-Efficient Personalized Federated Learning
via Decentralized Sparse Training [84.81043932706375]
We propose a novel personalized federated learning framework in a decentralized (peer-to-peer) communication protocol named Dis-PFL.
Dis-PFL employs personalized sparse masks to customize sparse local models on the edge.
We demonstrate that our method can easily adapt to heterogeneous local clients with varying computation complexities.
arXiv Detail & Related papers (2022-06-01T02:20:57Z) - SLAM: A Unified Encoder for Speech and Language Modeling via Speech-Text
Joint Pre-Training [33.02912456062474]
We build a single encoder with the BERT objective on unlabeled text together with the w2v-BERT objective on unlabeled speech.
We demonstrate that incorporating both speech and text data during pre-training can significantly improve downstream quality on CoVoST2 speech translation.
arXiv Detail & Related papers (2021-10-20T00:59:36Z) - ATCSpeechNet: A multilingual end-to-end speech recognition framework for
air traffic control systems [15.527854608553824]
ATCSpeechNet is proposed to tackle the issue of translating communication speech into human-readable text in air traffic control systems.
An end-to-end paradigm is developed to convert speech waveform into text directly, without any feature engineering or lexicon.
Experimental results on the ATCSpeech corpus demonstrate that the proposed approach achieves a high performance with a very small labeled corpus.
arXiv Detail & Related papers (2021-02-17T02:27:09Z) - Adaptive Prototypical Networks with Label Words and Joint Representation
Learning for Few-Shot Relation Classification [17.237331828747006]
This work focuses on few-shot relation classification (FSRC)
We propose an adaptive mixture mechanism to add label words to the representation of the class prototype.
Experiments have been conducted on FewRel under different few-shot (FS) settings.
arXiv Detail & Related papers (2021-01-10T11:25:42Z) - Federated Unsupervised Representation Learning [56.715917111878106]
We formulate a new problem in federated learning called Federated Unsupervised Representation Learning (FURL) to learn a common representation model without supervision.
FedCA is composed of two key modules: dictionary module to aggregate the representations of samples from each client and share with all clients for consistency of representation space and alignment module to align the representation of each client on a base model trained on a public data.
arXiv Detail & Related papers (2020-10-18T13:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.