Taking Notes on the Fly Helps BERT Pre-training
- URL: http://arxiv.org/abs/2008.01466v2
- Date: Sun, 14 Mar 2021 15:37:11 GMT
- Title: Taking Notes on the Fly Helps BERT Pre-training
- Authors: Qiyu Wu, Chen Xing, Yatao Li, Guolin Ke, Di He, Tie-Yan Liu
- Abstract summary: Taking Notes on the Fly (TNF) takes notes for rare words on the fly during pre-training to help the model understand them when they occur next time.
TNF provides better data utilization since cross-sentence information is employed to cover the inadequate semantics caused by rare words in the sentences.
- Score: 94.43953312613577
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: How to make unsupervised language pre-training more efficient and less
resource-intensive is an important research direction in NLP. In this paper, we
focus on improving the efficiency of language pre-training methods through
providing better data utilization. It is well-known that in language data
corpus, words follow a heavy-tail distribution. A large proportion of words
appear only very few times and the embeddings of rare words are usually poorly
optimized. We argue that such embeddings carry inadequate semantic signals,
which could make the data utilization inefficient and slow down the
pre-training of the entire model. To mitigate this problem, we propose Taking
Notes on the Fly (TNF), which takes notes for rare words on the fly during
pre-training to help the model understand them when they occur next time.
Specifically, TNF maintains a note dictionary and saves a rare word's
contextual information in it as notes when the rare word occurs in a sentence.
When the same rare word occurs again during training, the note information
saved beforehand can be employed to enhance the semantics of the current
sentence. By doing so, TNF provides better data utilization since
cross-sentence information is employed to cover the inadequate semantics caused
by rare words in the sentences. We implement TNF on both BERT and ELECTRA to
check its efficiency and effectiveness. Experimental results show that TNF's
training time is $60\%$ less than its backbone pre-training models when
reaching the same performance. When trained with the same number of iterations,
TNF outperforms its backbone methods on most of downstream tasks and the
average GLUE score. Source code is attached in the supplementary material.
Related papers
- Enhancing Medical Specialty Assignment to Patients using NLP Techniques [0.0]
We propose an alternative approach that achieves superior performance while being computationally efficient.
Specifically, we utilize keywords to train a deep learning architecture that outperforms a language model pretrained on a large corpus of text.
Our results demonstrate that utilizing keywords for text classification significantly improves classification performance.
arXiv Detail & Related papers (2023-12-09T14:13:45Z) - Revisit Few-shot Intent Classification with PLMs: Direct Fine-tuning vs. Continual Pre-training [20.98770732015944]
Few-shot intent detection involves training a deep learning model to classify utterances based on their underlying intents using only a small amount of labeled data.
We show that continual pre-training may not be essential, since the overfitting problem of PLMs on this task may not be as serious as expected.
To maximize the utilization of the limited available data, we propose a context augmentation method and leverage sequential self-distillation to boost performance.
arXiv Detail & Related papers (2023-06-08T15:26:52Z) - Towards preserving word order importance through Forced Invalidation [80.33036864442182]
We show that pre-trained language models are insensitive to word order.
We propose Forced Invalidation to help preserve the importance of word order.
Our experiments demonstrate that Forced Invalidation significantly improves the sensitivity of the models to word order.
arXiv Detail & Related papers (2023-04-11T13:42:10Z) - Curriculum-Based Self-Training Makes Better Few-Shot Learners for
Data-to-Text Generation [56.98033565736974]
We propose Curriculum-Based Self-Training (CBST) to leverage unlabeled data in a rearranged order determined by the difficulty of text generation.
Our method can outperform fine-tuning and task-adaptive pre-training methods, and achieve state-of-the-art performance in the few-shot setting of data-to-text generation.
arXiv Detail & Related papers (2022-06-06T16:11:58Z) - Sample Efficient Approaches for Idiomaticity Detection [6.481818246474555]
This work explores sample efficient methods of idiomaticity detection.
In particular, we study the impact of Pattern Exploit Training (PET), a few-shot method of classification, and BERTRAM, an efficient method of creating contextual embeddings.
Our experiments show that whilePET improves performance on English, they are much less effective on Portuguese and Galician, leading to an overall performance about on par with vanilla mBERT.
arXiv Detail & Related papers (2022-05-23T13:46:35Z) - How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial
Robustness? [121.57551065856164]
We propose Robust Informative Fine-Tuning (RIFT) as a novel adversarial fine-tuning method from an information-theoretical perspective.
RIFT encourages an objective model to retain the features learned from the pre-trained model throughout the entire fine-tuning process.
Experimental results show that RIFT consistently outperforms the state-of-the-arts on two popular NLP tasks.
arXiv Detail & Related papers (2021-12-22T05:04:41Z) - Rejuvenating Low-Frequency Words: Making the Most of Parallel Data in
Non-Autoregressive Translation [98.11249019844281]
Knowledge distillation (KD) is commonly used to construct synthetic data for training non-autoregressive translation (NAT) models.
We propose reverse KD to rejuvenate more alignments for low-frequency target words.
Results demonstrate that the proposed approach can significantly and universally improve translation quality.
arXiv Detail & Related papers (2021-06-02T02:41:40Z) - Improving accuracy of rare words for RNN-Transducer through unigram
shallow fusion [9.071295269523068]
We propose unigram shallow fusion (USF) to improve rare words for RNN-T.
We show that this simple method can improve performance on rare words by 3.7% WER relative without degradation on general test set.
arXiv Detail & Related papers (2020-11-30T22:06:02Z) - GiBERT: Introducing Linguistic Knowledge into BERT through a Lightweight
Gated Injection Method [29.352569563032056]
We propose a novel method to explicitly inject linguistic knowledge in the form of word embeddings into a pre-trained BERT.
Our performance improvements on multiple semantic similarity datasets when injecting dependency-based and counter-fitted embeddings indicate that such information is beneficial and currently missing from the original model.
arXiv Detail & Related papers (2020-10-23T17:00:26Z) - Improving Text Generation with Student-Forcing Optimal Transport [122.11881937642401]
We propose using optimal transport (OT) to match the sequences generated in training and testing modes.
An extension is also proposed to improve the OT learning, based on the structural and contextual information of the text sequences.
The effectiveness of the proposed method is validated on machine translation, text summarization, and text generation tasks.
arXiv Detail & Related papers (2020-10-12T19:42:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.