What and Where: Learn to Plug Adapters via NAS for Multi-Domain Learning
- URL: http://arxiv.org/abs/2007.12415v2
- Date: Tue, 18 May 2021 13:54:27 GMT
- Title: What and Where: Learn to Plug Adapters via NAS for Multi-Domain Learning
- Authors: Hanbin Zhao, Hao Zeng, Xin Qin, Yongjian Fu, Hui Wang, Bourahla Omar,
and Xi Li
- Abstract summary: We propose a data-driven adapter plugging strategy with Neural Architecture Search (NAS), which automatically determines where to plug for those adapter modules.
We also propose a NAS-adapter module for adapter structure design in a NAS-driven learning scheme, which automatically discovers effective adapter module structures for different domains.
- Score: 15.204498046112194
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As an important and challenging problem, multi-domain learning (MDL)
typically seeks for a set of effective lightweight domain-specific adapter
modules plugged into a common domain-agnostic network. Usually, existing ways
of adapter plugging and structure design are handcrafted and fixed for all
domains before model learning, resulting in the learning inflexibility and
computational intensiveness. With this motivation, we propose to learn a
data-driven adapter plugging strategy with Neural Architecture Search (NAS),
which automatically determines where to plug for those adapter modules.
Furthermore, we propose a NAS-adapter module for adapter structure design in a
NAS-driven learning scheme, which automatically discovers effective adapter
module structures for different domains. Experimental results demonstrate the
effectiveness of our MDL model against existing approaches under the conditions
of comparable performance.
Related papers
- Plug-and-Play Transformer Modules for Test-Time Adaptation [54.80435317208111]
We introduce PLUTO: a Plug-and-pLay modUlar Test-time domain adaptatiOn strategy.
We pre-train a large set of modules, each specialized for different source domains.
We harness multiple most-relevant source domains in a single inference call.
arXiv Detail & Related papers (2024-01-06T00:24:50Z) - Adapters: A Unified Library for Parameter-Efficient and Modular Transfer
Learning [109.25673110120906]
We introduce Adapters, an open-source library that unifies parameter-efficient and modular transfer learning in large language models.
By integrating 10 diverse adapter methods into a unified interface, Adapters offers ease of use and flexible configuration.
arXiv Detail & Related papers (2023-11-18T13:53:26Z) - Divide and Adapt: Active Domain Adaptation via Customized Learning [56.79144758380419]
We present Divide-and-Adapt (DiaNA), a new ADA framework that partitions the target instances into four categories with stratified transferable properties.
With a novel data subdivision protocol based on uncertainty and domainness, DiaNA can accurately recognize the most gainful samples.
Thanks to the "divideand-adapt" spirit, DiaNA can handle data with large variations of domain gap.
arXiv Detail & Related papers (2023-07-21T14:37:17Z) - AdapterSoup: Weight Averaging to Improve Generalization of Pretrained
Language Models [127.04370753583261]
Pretrained language models (PLMs) are trained on massive corpora, but often need to specialize to specific domains.
A solution is to use a related-domain adapter for the novel domain at test time.
We introduce AdapterSoup, an approach that performs weight-space averaging of adapters trained on different domains.
arXiv Detail & Related papers (2023-02-14T13:09:23Z) - Neural Attentive Circuits [93.95502541529115]
We introduce a general purpose, yet modular neural architecture called Neural Attentive Circuits (NACs)
NACs learn the parameterization and a sparse connectivity of neural modules without using domain knowledge.
NACs achieve an 8x speedup at inference time while losing less than 3% performance.
arXiv Detail & Related papers (2022-10-14T18:00:07Z) - Adaptable Adapters [74.65986170056945]
State-of-the-art pretrained NLP models contain a hundred million to trillion parameters.
Adaptable adapters contain different activation functions for different layers and different input data.
We show that adaptable adapters achieve on-par performances with the standard adapter architecture.
arXiv Detail & Related papers (2022-05-03T14:59:27Z) - Model-agnostic Multi-Domain Learning with Domain-Specific Adapters for
Action Recognition [0.0]
The proposed method inserts domain-specific adapters between layers of domain-independent layers of a backbone network.
Unlike a multi-head network that switches classification heads only, our model switches not only the heads, but also the adapters for facilitating to learn feature representations universal to multiple domains.
arXiv Detail & Related papers (2022-04-15T00:02:13Z) - AdapterHub Playground: Simple and Flexible Few-Shot Learning with
Adapters [34.86139827292556]
Open-access dissemination of pretrained language models has led to a democratization of state-of-the-art natural language processing (NLP) research.
This also allows people outside of NLP to use such models and adapt them to specific use-cases.
In this work, we aim to overcome this gap by providing a tool which allows researchers to leverage pretrained models without writing a single line of code.
arXiv Detail & Related papers (2021-08-18T11:56:01Z) - Network Architecture Search for Domain Adaptation [11.24426822697648]
We present Neural Architecture Search for Domain Adaptation (NASDA), a principle framework that leverages differentiable neural architecture search to derive the optimal network architecture for domain adaptation task.
We demonstrate experimentally that NASDA leads to state-of-the-art performance on several domain adaptation benchmarks.
arXiv Detail & Related papers (2020-08-13T06:15:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.