FedAdapter: Efficient Federated Learning for Modern NLP
- URL: http://arxiv.org/abs/2205.10162v2
- Date: Mon, 8 May 2023 19:50:56 GMT
- Title: FedAdapter: Efficient Federated Learning for Modern NLP
- Authors: Dongqi Cai, Yaozong Wu, Shangguang Wang, Felix Xiaozhu Lin, Mengwei Xu
- Abstract summary: Fine-tuning pre-trained models for downstream tasks often requires private data.
FedNLP is prohibitively slow due to the large model sizes and the resultant high network/computation cost.
We propose FedAdapter, a framework that enhances the existing FedNLP with two novel designs.
- Score: 2.6706511009396023
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Transformer-based pre-trained models have revolutionized NLP for superior
performance and generality. Fine-tuning pre-trained models for downstream tasks
often requires private data, for which federated learning is the de-facto
approach (i.e., FedNLP). However, our measurements show that FedNLP is
prohibitively slow due to the large model sizes and the resultant high
network/computation cost. Towards practical FedNLP, we identify as the key
building blocks adapters, small bottleneck modules inserted at a variety of
model layers. A key challenge is to properly configure the depth and width of
adapters, to which the training speed and efficiency is highly sensitive. No
silver-bullet configuration exists: the optimal choice varies across downstream
NLP tasks, desired model accuracy, and mobile resources. To automate adapter
configuration, we propose FedAdapter, a framework that enhances the existing
FedNLP with two novel designs. First, FedAdapter progressively upgrades the
adapter configuration throughout a training session; the principle is to
quickly learn shallow knowledge by only training fewer and smaller adapters at
the model's top layers, and incrementally learn deep knowledge by incorporating
deeper and larger adapters. Second, FedAdapter continuously profiles future
adapter configurations by allocating participant devices to trial groups.
Extensive experiments show that FedAdapter can reduce FedNLP's model
convergence delay to no more than several hours, which is up to 155.5$\times$
faster compared to vanilla FedNLP and 48$\times$ faster compared to strong
baselines.
Related papers
- Improving Robustness of Foundation Models in Domain Adaptation with Soup-Adapters [0.0]
We show that by training multiple independent adapters and averaging their outputs, the new model has a higher performance and is more robust to distribution shifts compared to any individual adapter.<n>This is also the first study to explore CLIP adapter-style techniques for DINOv2 and to directly compare them with CLIP in this setting.
arXiv Detail & Related papers (2025-07-08T09:26:10Z) - Adaptive Adapter Routing for Long-Tailed Class-Incremental Learning [55.384428765798496]
New data exhibits a long-tailed distribution, such as e-commerce platform reviews.
This necessitates continuous model learning imbalanced data without forgetting.
We introduce AdaPtive Adapter RouTing (APART) as an exemplar-free solution for LTCIL.
arXiv Detail & Related papers (2024-09-11T17:52:00Z) - MerA: Merging Pretrained Adapters For Few-Shot Learning [71.44422347502409]
We propose textbftextttMerging Pretrained Adapters (MerA) that efficiently incorporates pretrained adapters to a single model through model fusion.
Experiments on two PLMs demonstrate that MerA substantial improvements compared to both single adapters and AdapterFusion.
arXiv Detail & Related papers (2023-08-30T12:10:17Z) - Adaptable Adapters [74.65986170056945]
State-of-the-art pretrained NLP models contain a hundred million to trillion parameters.
Adaptable adapters contain different activation functions for different layers and different input data.
We show that adaptable adapters achieve on-par performances with the standard adapter architecture.
arXiv Detail & Related papers (2022-05-03T14:59:27Z) - Unidirectional Thin Adapter for Efficient Adaptation of Deep Neural
Networks [5.995023738151625]
We propose a new adapter network for adapting a pre-trained deep neural network to a target domain with minimal computation.
The proposed model, unidirectional thin adapter (UDTA), helps the classifier adapt to new data by providing auxiliary features that complement the backbone network.
In experiments on five fine-grained classification datasets, UDTA significantly reduced computation and training time required for backpropagation.
arXiv Detail & Related papers (2022-03-20T06:06:43Z) - Tip-Adapter: Training-free CLIP-Adapter for Better Vision-Language
Modeling [78.62723847797382]
We propose textbfTraining-Free CLtextbfIP-textbfAdapter (textbfTip-Adapter), which not only inherits CLIP's training-free advantage but also performs comparably or even better than CLIP-Adapter.
We conduct extensive experiments of few-shot classification on ImageNet and other 10 datasets to demonstrate the superiority of proposed Tip-Adapter.
arXiv Detail & Related papers (2021-11-06T18:09:22Z) - FedAdapt: Adaptive Offloading for IoT Devices in Federated Learning [2.5775113252104216]
Federated Learning (FL) on Internet-of-Things devices is necessitated by the large volumes of data they produce and growing concerns of data privacy.
This paper presents FedAdapt, an adaptive offloading FL framework to mitigate the aforementioned challenges.
arXiv Detail & Related papers (2021-07-09T07:29:55Z) - AdapterDrop: On the Efficiency of Adapters in Transformers [53.845909603631945]
Massively pre-trained transformer models are computationally expensive to fine-tune, slow for inference, and have large storage requirements.
Recent approaches tackle these shortcomings by training smaller models, dynamically reducing the model size, and by training light-weight adapters.
arXiv Detail & Related papers (2020-10-22T17:49:42Z) - AdapterHub: A Framework for Adapting Transformers [148.6877231725939]
AdapterHub is a framework that allows dynamic "stitching-in" of pre-trained adapters for different tasks and languages.
Our framework enables scalable and easy access to sharing of task-specific models.
arXiv Detail & Related papers (2020-07-15T15:56:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.