BERT-JAM: Boosting BERT-Enhanced Neural Machine Translation with Joint
Attention
- URL: http://arxiv.org/abs/2011.04266v1
- Date: Mon, 9 Nov 2020 09:30:37 GMT
- Title: BERT-JAM: Boosting BERT-Enhanced Neural Machine Translation with Joint
Attention
- Authors: Zhebin Zhang, Sai Wu, Dawei Jiang, Gang Chen
- Abstract summary: We propose a novel BERT-enhanced neural machine translation model called BERT-JAM.
BERT-JAM uses joint-attention modules to allow the encoder/decoder layers to dynamically allocate attention between different representations.
Our experiments show that BERT-JAM achieves SOTA BLEU scores on multiple translation tasks.
- Score: 9.366359346271567
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: BERT-enhanced neural machine translation (NMT) aims at leveraging
BERT-encoded representations for translation tasks. A recently proposed
approach uses attention mechanisms to fuse Transformer's encoder and decoder
layers with BERT's last-layer representation and shows enhanced performance.
However, their method doesn't allow for the flexible distribution of attention
between the BERT representation and the encoder/decoder representation. In this
work, we propose a novel BERT-enhanced NMT model called BERT-JAM which improves
upon existing models from two aspects: 1) BERT-JAM uses joint-attention modules
to allow the encoder/decoder layers to dynamically allocate attention between
different representations, and 2) BERT-JAM allows the encoder/decoder layers to
make use of BERT's intermediate representations by composing them using a gated
linear unit (GLU). We train BERT-JAM with a novel three-phase optimization
strategy that progressively unfreezes different components of BERT-JAM. Our
experiments show that BERT-JAM achieves SOTA BLEU scores on multiple
translation tasks.
Related papers
- Adapted Multimodal BERT with Layer-wise Fusion for Sentiment Analysis [84.12658971655253]
We propose Adapted Multimodal BERT, a BERT-based architecture for multimodal tasks.
adapter adjusts the pretrained language model for the task at hand, while the fusion layers perform task-specific, layer-wise fusion of audio-visual information with textual BERT representations.
In our ablations we see that this approach leads to efficient models, that can outperform their fine-tuned counterparts and are robust to input noise.
arXiv Detail & Related papers (2022-12-01T17:31:42Z) - BERTVision -- A Parameter-Efficient Approach for Question Answering [0.0]
We present a highly parameter efficient approach for Question Answering that significantly reduces the need for extended BERT fine-tuning.
Our method uses information from the hidden state activations of each BERT transformer layer, which is discarded during typical BERT inference.
Our experiments show that this approach works well not only for span QA, but also for classification, suggesting that it may be to a wider range of tasks.
arXiv Detail & Related papers (2022-02-24T17:16:25Z) - PromptBERT: Improving BERT Sentence Embeddings with Prompts [95.45347849834765]
We propose a prompt based sentence embeddings method which can reduce token embeddings biases and make the original BERT layers more effective.
We also propose a novel unsupervised training objective by the technology of template denoising, which substantially shortens the performance gap between the supervised and unsupervised setting.
Our fine-tuned method outperforms the state-of-the-art method SimCSE in both unsupervised and supervised settings.
arXiv Detail & Related papers (2022-01-12T06:54:21Z) - Incorporating BERT into Parallel Sequence Decoding with Adapters [82.65608966202396]
We propose to take two different BERT models as the encoder and decoder respectively, and fine-tune them by introducing simple and lightweight adapter modules.
We obtain a flexible and efficient model which is able to jointly leverage the information contained in the source-side and target-side BERT models.
Our framework is based on a parallel sequence decoding algorithm named Mask-Predict considering the bi-directional and conditional independent nature of BERT.
arXiv Detail & Related papers (2020-10-13T03:25:15Z) - CERT: Contrastive Self-supervised Learning for Language Understanding [20.17416958052909]
We propose CERT: Contrastive self-supervised Representations from Transformers.
CERT pretrains language representation models using contrastive self-supervised learning at the sentence level.
We evaluate CERT on 11 natural language understanding tasks in the GLUE benchmark where CERT outperforms BERT on 7 tasks, achieves the same performance as BERT on 2 tasks, and performs worse than BERT on 2 tasks.
arXiv Detail & Related papers (2020-05-16T16:20:38Z) - DynaBERT: Dynamic BERT with Adaptive Width and Depth [55.18269622415814]
We propose a novel dynamic BERT model (abbreviated as DynaBERT)
It can flexibly adjust the size and latency by selecting adaptive width and depth.
It consistently outperforms existing BERT compression methods.
arXiv Detail & Related papers (2020-04-08T15:06:28Z) - Incorporating BERT into Neural Machine Translation [251.54280200353674]
We propose a new algorithm named BERT-fused model, in which we first use BERT to extract representations for an input sequence.
We conduct experiments on supervised (including sentence-level and document-level translations), semi-supervised and unsupervised machine translation, and achieve state-of-the-art results on seven benchmark datasets.
arXiv Detail & Related papers (2020-02-17T08:13:36Z) - BERT's output layer recognizes all hidden layers? Some Intriguing
Phenomena and a simple way to boost BERT [53.63288887672302]
Bidirectional Representations from Transformers (BERT) have achieved tremendous success in many natural language processing (NLP) tasks.
We find that surprisingly the output layer of BERT can reconstruct the input sentence by directly taking each layer of BERT as input.
We propose a quite simple method to boost the performance of BERT.
arXiv Detail & Related papers (2020-01-25T13:35:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.