FedNLP: A Research Platform for Federated Learning in Natural Language
Processing
- URL: http://arxiv.org/abs/2104.08815v1
- Date: Sun, 18 Apr 2021 11:04:49 GMT
- Title: FedNLP: A Research Platform for Federated Learning in Natural Language
Processing
- Authors: Bill Yuchen Lin, Chaoyang He, Zihang Zeng, Hulin Wang, Yufen Huang,
Mahdi Soltanolkotabi, Xiang Ren, Salman Avestimehr
- Abstract summary: We present the FedNLP, a research platform for federated learning in NLP.
FedNLP supports various popular task formulations in NLP such as text classification, sequence tagging, question answering, seq2seq generation, and language modeling.
Preliminary experiments with FedNLP reveal that there exists a large performance gap between learning on decentralized and centralized datasets.
- Score: 55.01246123092445
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Increasing concerns and regulations about data privacy, necessitate the study
of privacy-preserving methods for natural language processing (NLP)
applications. Federated learning (FL) provides promising methods for a large
number of clients (i.e., personal devices or organizations) to collaboratively
learn a shared global model to benefit all clients, while allowing users to
keep their data locally. To facilitate FL research in NLP, we present the
FedNLP, a research platform for federated learning in NLP. FedNLP supports
various popular task formulations in NLP such as text classification, sequence
tagging, question answering, seq2seq generation, and language modeling. We also
implement an interface between Transformer language models (e.g., BERT) and FL
methods (e.g., FedAvg, FedOpt, etc.) for distributed training. The evaluation
protocol of this interface supports a comprehensive collection of non-IID
partitioning strategies. Our preliminary experiments with FedNLP reveal that
there exists a large performance gap between learning on decentralized and
centralized datasets -- opening intriguing and exciting future research
directions aimed at developing FL methods suited to NLP tasks.
Related papers
- Natural Language Processing for Dialects of a Language: A Survey [56.93337350526933]
State-of-the-art natural language processing (NLP) models are trained on massive training corpora, and report a superlative performance on evaluation datasets.
This survey delves into an important attribute of these datasets: the dialect of a language.
Motivated by the performance degradation of NLP models for dialectic datasets and its implications for the equity of language technologies, we survey past research in NLP for dialects in terms of datasets, and approaches.
arXiv Detail & Related papers (2024-01-11T03:04:38Z) - Tunable Soft Prompts are Messengers in Federated Learning [55.924749085481544]
Federated learning (FL) enables multiple participants to collaboratively train machine learning models using decentralized data sources.
The lack of model privacy protection in FL becomes an unneglectable challenge.
We propose a novel FL training approach that accomplishes information exchange among participants via tunable soft prompts.
arXiv Detail & Related papers (2023-11-12T11:01:10Z) - DPP-based Client Selection for Federated Learning with Non-IID Data [97.1195165400568]
This paper proposes a client selection (CS) method to tackle the communication bottleneck of federated learning (FL)
We first analyze the effect of CS in FL and show that FL training can be accelerated by adequately choosing participants to diversify the training dataset in each round of training.
We leverage data profiling and determinantal point process (DPP) sampling techniques to develop an algorithm termed Federated Learning with DPP-based Participant Selection (FL-DP$3$S)
arXiv Detail & Related papers (2023-03-30T13:14:54Z) - HugNLP: A Unified and Comprehensive Library for Natural Language
Processing [14.305751154503133]
We introduce HugNLP, a library for natural language processing (NLP) with the prevalent backend of HuggingFace Transformers.
HugNLP consists of a hierarchical structure including models, processors and applications that unifies the learning process of pre-trained language models (PLMs) on different NLP tasks.
arXiv Detail & Related papers (2023-02-28T03:38:26Z) - Collaborating Heterogeneous Natural Language Processing Tasks via
Federated Learning [55.99444047920231]
The proposed ATC framework achieves significant improvements compared with various baseline methods.
We conduct extensive experiments on six widely-used datasets covering both Natural Language Understanding (NLU) and Natural Language Generation (NLG) tasks.
arXiv Detail & Related papers (2022-12-12T09:27:50Z) - Pretrained Models for Multilingual Federated Learning [38.19507070702635]
We study how multilingual text impacts Federated Learning (FL) algorithms.
We explore three multilingual language tasks, language modeling, machine translation, and text classification.
Our results show that using pretrained models reduces the negative effects of FL, helping them to perform near or better than centralized (no privacy) learning.
arXiv Detail & Related papers (2022-06-06T00:20:30Z) - Meta Learning for Natural Language Processing: A Survey [88.58260839196019]
Deep learning has been the mainstream technique in natural language processing (NLP) area.
Deep learning requires many labeled data and is less generalizable across domains.
Meta-learning is an arising field in machine learning studying approaches to learn better algorithms.
arXiv Detail & Related papers (2022-05-03T13:58:38Z) - Federated Learning Meets Natural Language Processing: A Survey [12.224792145700562]
Federated Learning aims to learn machine learning models from multiple decentralized edge devices (e.g. mobiles) or servers without sacrificing local data privacy.
Recent Natural Language Processing techniques rely on deep learning and large pre-trained language models.
arXiv Detail & Related papers (2021-07-27T05:07:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.