Learning and Analyzing Generation Order for Undirected Sequence Models
- URL: http://arxiv.org/abs/2112.09097v1
- Date: Thu, 16 Dec 2021 18:29:07 GMT
- Title: Learning and Analyzing Generation Order for Undirected Sequence Models
- Authors: Yichen Jiang, Mohit Bansal
- Abstract summary: We train a policy that learns the generation order for a pre-trained, undirected translation model via reinforcement learning.
We show that the translations by our learned orders achieve higher BLEU scores than the outputs decoded from left to right or decoded by the learned order from Mansimov et al.
Our findings could provide more insights on the mechanism of undirected generation models and encourage further research in this direction.
- Score: 86.10875837475783
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Undirected neural sequence models have achieved performance competitive with
the state-of-the-art directed sequence models that generate monotonically from
left to right in machine translation tasks. In this work, we train a policy
that learns the generation order for a pre-trained, undirected translation
model via reinforcement learning. We show that the translations decoded by our
learned orders achieve higher BLEU scores than the outputs decoded from left to
right or decoded by the learned order from Mansimov et al. (2019) on the WMT'14
German-English translation task. On examples with a maximum source and target
length of 30 from De-En, WMT'16 English-Romanian, and WMT'21 English-Chinese
translation tasks, our learned order outperforms all heuristic generation
orders on four out of six tasks. We next carefully analyze the learned order
patterns via qualitative and quantitative analysis. We show that our policy
generally follows an outer-to-inner order, predicting the left-most and
right-most positions first, and then moving toward the middle while skipping
less important words at the beginning. Furthermore, the policy usually predicts
positions for a single syntactic constituent structure in consecutive steps. We
believe our findings could provide more insights on the mechanism of undirected
generation models and encourage further research in this direction. Our code is
publicly available at https://github.com/jiangycTarheel/undirected-generation
Related papers
- Learning to Predict Concept Ordering for Common Sense Generation [32.2052248473022]
We study the relationship between the ordering of the input concepts and the quality of the generated sentences.
We find that BART-large model consistently outperforms all other LMs considered in this study.
The larger GPT3-based large language models (LLMs) variants do not necessarily outperform much smaller LMs on this task.
arXiv Detail & Related papers (2023-09-12T16:27:18Z) - Instruction Position Matters in Sequence Generation with Large Language
Models [67.87516654892343]
Large language models (LLMs) are capable of performing conditional sequence generation tasks, such as translation or summarization.
We propose enhancing the instruction-following capability of LLMs by shifting the position of task instructions after the input sentences.
arXiv Detail & Related papers (2023-08-23T12:36:57Z) - Few-shot Subgoal Planning with Language Models [58.11102061150875]
We show that language priors encoded in pre-trained language models allow us to infer fine-grained subgoal sequences.
In contrast to recent methods which make strong assumptions about subgoal supervision, our experiments show that language models can infer detailed subgoal sequences without any fine-tuning.
arXiv Detail & Related papers (2022-05-28T01:03:30Z) - Summarize and Generate to Back-translate: Unsupervised Translation of
Programming Languages [86.08359401867577]
Back-translation is widely known for its effectiveness for neural machine translation when little to no parallel data is available.
We propose performing back-translation via code summarization and generation.
We show that our proposed approach performs competitively with state-of-the-art methods.
arXiv Detail & Related papers (2022-05-23T08:20:41Z) - Discovering Non-monotonic Autoregressive Orderings with Variational
Inference [67.27561153666211]
We develop an unsupervised parallelizable learner that discovers high-quality generation orders purely from training data.
We implement the encoder as a Transformer with non-causal attention that outputs permutations in one forward pass.
Empirical results in language modeling tasks demonstrate that our method is context-aware and discovers orderings that are competitive with or even better than fixed orders.
arXiv Detail & Related papers (2021-10-27T16:08:09Z) - COCO-LM: Correcting and Contrasting Text Sequences for Language Model
Pretraining [59.169836983883656]
COCO-LM is a new self-supervised learning framework that pretrains Language Models by COrrecting challenging errors and COntrasting text sequences.
COCO-LM employs an auxiliary language model to mask-and-predict tokens in original text sequences.
Our analyses reveal that COCO-LM's advantages come from its challenging training signals, more contextualized token representations, and regularized sequence representations.
arXiv Detail & Related papers (2021-02-16T22:24:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.