Calibrating Sequence likelihood Improves Conditional Language Generation
- URL: http://arxiv.org/abs/2210.00045v1
- Date: Fri, 30 Sep 2022 19:16:16 GMT
- Title: Calibrating Sequence likelihood Improves Conditional Language Generation
- Authors: Yao Zhao, Misha Khalman, Rishabh Joshi, Shashi Narayan, Mohammad
Saleh, Peter J. Liu
- Abstract summary: Conditional language models are predominantly trained with maximum likelihood estimation (MLE)
While MLE trained models assign high probability to plausible sequences given the context, the model probabilities often do not accurately rank-order generated sequences by quality.
We introduce sequence likelihood calibration (SLiC) where the likelihood of model generated sequences are calibrated to better align with reference sequences in the model's latent space.
- Score: 39.35161650538767
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conditional language models are predominantly trained with maximum likelihood
estimation (MLE), giving probability mass to sparsely observed target
sequences. While MLE trained models assign high probability to plausible
sequences given the context, the model probabilities often do not accurately
rank-order generated sequences by quality. This has been empirically observed
in beam search decoding as output quality degrading with large beam sizes, and
decoding strategies benefiting from heuristics such as length normalization and
repetition-blocking. In this work, we introduce sequence likelihood calibration
(SLiC) where the likelihood of model generated sequences are calibrated to
better align with reference sequences in the model's latent space. With SLiC,
decoding heuristics become unnecessary and decoding candidates' quality
significantly improves regardless of the decoding method. Furthermore, SLiC
shows no sign of diminishing returns with model scale, and presents alternative
ways to improve quality with limited training and inference budgets. With SLiC,
we exceed or match SOTA results on a wide range of generation tasks spanning
abstractive summarization, question generation, abstractive question answering
and data-to-text generation, even with modest-sized models.
Related papers
- Calibrating Likelihoods towards Consistency in Summarization Models [22.023863165579602]
We argue that the main reason for such behavior is that the summarization models trained with maximum likelihood objective assign high probability to plausible sequences given the context.
In this work, we solve this problem by calibrating the likelihood of model generated sequences to better align with a consistency metric measured by natural language inference (NLI) models.
arXiv Detail & Related papers (2023-10-12T23:17:56Z) - Instruction Position Matters in Sequence Generation with Large Language
Models [67.87516654892343]
Large language models (LLMs) are capable of performing conditional sequence generation tasks, such as translation or summarization.
We propose enhancing the instruction-following capability of LLMs by shifting the position of task instructions after the input sentences.
arXiv Detail & Related papers (2023-08-23T12:36:57Z) - SequenceMatch: Imitation Learning for Autoregressive Sequence Modelling with Backtracking [60.109453252858806]
A maximum-likelihood (MLE) objective does not match a downstream use-case of autoregressively generating high-quality sequences.
We formulate sequence generation as an imitation learning (IL) problem.
This allows us to minimize a variety of divergences between the distribution of sequences generated by an autoregressive model and sequences from a dataset.
Our resulting method, SequenceMatch, can be implemented without adversarial training or architectural changes.
arXiv Detail & Related papers (2023-06-08T17:59:58Z) - A Non-monotonic Self-terminating Language Model [62.93465126911921]
In this paper, we focus on the problem of non-terminating sequences resulting from an incomplete decoding algorithm.
We first define an incomplete probable decoding algorithm which includes greedy search, top-$k$ sampling, and nucleus sampling.
We then propose a non-monotonic self-terminating language model, which relaxes the constraint of monotonically increasing termination probability.
arXiv Detail & Related papers (2022-10-03T00:28:44Z) - Out-of-Distribution Detection and Selective Generation for Conditional
Language Models [40.15896981028647]
Conditional language models (CLMs) are predominantly trained to classify the next token in an output sequence.
We present a highly accurate and lightweight OOD detection method for CLMs.
We show how our method can be used under the common and realistic setting of distribution shift for selective generation of high-quality outputs.
arXiv Detail & Related papers (2022-09-30T16:17:11Z) - CodeRL: Mastering Code Generation through Pretrained Models and Deep
Reinforcement Learning [92.36705236706678]
"CodeRL" is a new framework for program synthesis tasks through pretrained LMs and deep reinforcement learning.
During inference, we introduce a new generation procedure with a critical sampling strategy.
For the model backbones, we extended the encoder-decoder architecture of CodeT5 with enhanced learning objectives.
arXiv Detail & Related papers (2022-07-05T02:42:15Z) - Energy-Based Models for Code Generation under Compilability Constraints [2.9176992922046923]
In this work, we pose the problem of learning to generate compilable code as constraint satisfaction.
We define an Energy-Based Model (EBM) representing a pre-trained generative model with an imposed constraint of generating only compilable sequences.
We then use the KL-Adaptive Distributional Policy Gradient algorithm to train a generative model approxing the EBM.
arXiv Detail & Related papers (2021-06-09T11:06:32Z) - Consistency of a Recurrent Language Model With Respect to Incomplete
Decoding [67.54760086239514]
We study the issue of receiving infinite-length sequences from a recurrent language model.
We propose two remedies which address inconsistency: consistent variants of top-k and nucleus sampling, and a self-terminating recurrent language model.
arXiv Detail & Related papers (2020-02-06T19:56:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.