Towards Non-task-specific Distillation of BERT via Sentence
Representation Approximation
- URL: http://arxiv.org/abs/2004.03097v1
- Date: Tue, 7 Apr 2020 03:03:00 GMT
- Title: Towards Non-task-specific Distillation of BERT via Sentence
Representation Approximation
- Authors: Bowen Wu, Huan Zhang, Mengyuan Li, Zongsheng Wang, Qihang Feng,
Junhong Huang, Baoxun Wang
- Abstract summary: We propose a sentence representation approximating oriented distillation framework that can distill the pre-trained BERT into a simple LSTM based model.
Our model is able to perform transfer learning via fine-tuning to adapt to any sentence-level downstream task.
The experimental results on multiple NLP tasks from the GLUE benchmark show that our approach outperforms other task-specific distillation methods.
- Score: 17.62309851473892
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, BERT has become an essential ingredient of various NLP deep models
due to its effectiveness and universal-usability. However, the online
deployment of BERT is often blocked by its large-scale parameters and high
computational cost. There are plenty of studies showing that the knowledge
distillation is efficient in transferring the knowledge from BERT into the
model with a smaller size of parameters. Nevertheless, current BERT
distillation approaches mainly focus on task-specified distillation, such
methodologies lead to the loss of the general semantic knowledge of BERT for
universal-usability. In this paper, we propose a sentence representation
approximating oriented distillation framework that can distill the pre-trained
BERT into a simple LSTM based model without specifying tasks. Consistent with
BERT, our distilled model is able to perform transfer learning via fine-tuning
to adapt to any sentence-level downstream task. Besides, our model can further
cooperate with task-specific distillation procedures. The experimental results
on multiple NLP tasks from the GLUE benchmark show that our approach
outperforms other task-specific distillation methods or even much larger
models, i.e., ELMO, with efficiency well-improved.
Related papers
- Improving Knowledge Distillation for BERT Models: Loss Functions,
Mapping Methods, and Weight Tuning [1.1510009152620668]
This project investigates and applies knowledge distillation for BERT model compression.
We explore various techniques to improve knowledge distillation, including experimentation with loss functions, transformer layer mapping methods, and tuning the weights of attention and representation loss.
The goal of this work is to improve the efficiency and effectiveness of knowledge distillation, enabling the development of more efficient and accurate models for a range of natural language processing tasks.
arXiv Detail & Related papers (2023-08-26T20:59:21Z) - MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided
Adaptation [68.30497162547768]
We propose MoEBERT, which uses a Mixture-of-Experts structure to increase model capacity and inference speed.
We validate the efficiency and effectiveness of MoEBERT on natural language understanding and question answering tasks.
arXiv Detail & Related papers (2022-04-15T23:19:37Z) - BiBERT: Accurate Fully Binarized BERT [69.35727280997617]
BiBERT is an accurate fully binarized BERT to eliminate the performance bottlenecks.
Our method yields impressive 56.3 times and 31.2 times saving on FLOPs and model size.
arXiv Detail & Related papers (2022-03-12T09:46:13Z) - BERTVision -- A Parameter-Efficient Approach for Question Answering [0.0]
We present a highly parameter efficient approach for Question Answering that significantly reduces the need for extended BERT fine-tuning.
Our method uses information from the hidden state activations of each BERT transformer layer, which is discarded during typical BERT inference.
Our experiments show that this approach works well not only for span QA, but also for classification, suggesting that it may be to a wider range of tasks.
arXiv Detail & Related papers (2022-02-24T17:16:25Z) - Automatic Mixed-Precision Quantization Search of BERT [62.65905462141319]
Pre-trained language models such as BERT have shown remarkable effectiveness in various natural language processing tasks.
These models usually contain millions of parameters, which prevents them from practical deployment on resource-constrained devices.
We propose an automatic mixed-precision quantization framework designed for BERT that can simultaneously conduct quantization and pruning in a subgroup-wise level.
arXiv Detail & Related papers (2021-12-30T06:32:47Z) - Why Can You Lay Off Heads? Investigating How BERT Heads Transfer [37.9520341259181]
The main goal of distillation is to create a task-agnostic pre-trained model that can be fine-tuned on downstream tasks without fine-tuning its full-sized version.
Despite the progress of distillation, to what degree and for what reason a task-agnostic model can be created from distillation has not been well studied.
This work focuses on analyzing the acceptable deduction when distillation for guiding the future distillation procedure.
arXiv Detail & Related papers (2021-06-14T02:27:47Z) - LRC-BERT: Latent-representation Contrastive Knowledge Distillation for
Natural Language Understanding [12.208166079145538]
We propose a knowledge distillation method LRC-BERT based on contrastive learning to fit the output of the intermediate layer from the angular distance aspect.
By verifying 8 datasets on the General Language Understanding Evaluation (GLUE) benchmark, the performance of the proposed LRC-BERT exceeds the existing state-of-the-art methods.
arXiv Detail & Related papers (2020-12-14T08:39:38Z) - Pre-trained Summarization Distillation [121.14806854092672]
Recent work on distilling BERT for classification and regression tasks shows strong performance using direct knowledge distillation.
Alternatively, machine translation practitioners distill using pseudo-labeling, where a small model is trained on the translations of a larger model.
A third, simpler approach is to'shrink and fine-tune' (SFT), which avoids any explicit distillation by copying parameters to a smaller student model and then fine-tuning.
arXiv Detail & Related papers (2020-10-24T23:15:43Z) - TernaryBERT: Distillation-aware Ultra-low Bit BERT [53.06741585060951]
We propose TernaryBERT, which ternarizes the weights in a fine-tuned BERT model.
Experiments on the GLUE benchmark and SQuAD show that our proposed TernaryBERT outperforms the other BERT quantization methods.
arXiv Detail & Related papers (2020-09-27T10:17:28Z) - Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation [84.64004917951547]
Fine-tuning pre-trained language models like BERT has become an effective way in NLP.
In this paper, we improve the fine-tuning of BERT with two effective mechanisms: self-ensemble and self-distillation.
arXiv Detail & Related papers (2020-02-24T16:17:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.