CEV-LM: Controlled Edit Vector Language Model for Shaping Natural
Language Generations
- URL: http://arxiv.org/abs/2402.14290v1
- Date: Thu, 22 Feb 2024 05:07:31 GMT
- Title: CEV-LM: Controlled Edit Vector Language Model for Shaping Natural
Language Generations
- Authors: Samraj Moorjani, Adit Krishnan, Hari Sundaram
- Abstract summary: We introduce CEV-LM - a lightweight, semi-autoregressive language model that utilizes constrained edit vectors to control three complementary metrics.
We study an extensive set of state-of-the-art CTG models and find that CEV-LM provides significantly more targeted and precise control of these three metrics.
- Score: 5.148810760938979
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As large-scale language models become the standard for text generation, there
is a greater need to tailor the generations to be more or less concise,
targeted, and informative, depending on the audience/application. Existing
control approaches primarily adjust the semantic (e.g., emotion, topics),
structural (e.g., syntax tree, parts-of-speech), and lexical (e.g.,
keyword/phrase inclusion) properties of text, but are insufficient to
accomplish complex objectives such as pacing which control the complexity and
readability of the text. In this paper, we introduce CEV-LM - a lightweight,
semi-autoregressive language model that utilizes constrained edit vectors to
control three complementary metrics (speed, volume, and circuitousness) that
quantify the shape of text (e.g., pacing of content). We study an extensive set
of state-of-the-art CTG models and find that CEV-LM provides significantly more
targeted and precise control of these three metrics while preserving semantic
content, using less training data, and containing fewer parameters.
Related papers
- Controllable Text Generation for Large Language Models: A Survey [27.110528099257156]
This paper systematically reviews the latest advancements in Controllable Text Generation for Large Language Models.
We categorize CTG tasks into two primary types: content control and control.
We address key challenges in current research, including reduced fluency and practicality.
arXiv Detail & Related papers (2024-08-22T17:59:04Z) - Controlled Text Generation via Language Model Arithmetic [7.687678490751105]
We introduce model arithmetic, a novel inference framework for composing and biasing Large Language Models.
We show that model arithmetic allows fine-grained control of generated text while outperforming state-of-the-art on the task of toxicity reduction.
arXiv Detail & Related papers (2023-11-24T13:41:12Z) - Controlled Text Generation for Black-box Language Models via Score-based Progressive Editor [32.913066883646074]
Controlled text generation is very important for the practical use of language models.
Existing methods are inapplicable to black-box models or suffer a significant trade-off between controlling the generated text and maintaining its fluency.
This paper introduces the Score-based Progressive Editor (ScoPE), a novel approach designed to overcome these issues.
arXiv Detail & Related papers (2023-11-13T16:03:23Z) - Successor Features for Efficient Multisubject Controlled Text Generation [48.37713738712319]
We introduce SF-GEN, which is grounded in two primary concepts: successor features (SFs) and language model rectification.
SF-GEN seamlessly integrates the two to enable dynamic steering of text generation with no need to alter the LLM's parameters.
To the best of our knowledge, our research represents the first application of successor features in text generation.
arXiv Detail & Related papers (2023-11-03T00:17:08Z) - Adapting Language Models to Compress Contexts [71.98287002918941]
Transformer-based language models (LMs) are powerful and widely-applicable tools, but their usefulness is constrained by a finite context window.
We propose to adapt pre-trained LMs into AutoCompressors, which are capable of compressing long contexts into compact summary vectors.
We fine-tune OPT and Llama-2 models on sequences of up to 30,720 tokens and show that AutoCompressors can utilize long contexts to improve perplexity.
arXiv Detail & Related papers (2023-05-24T06:42:44Z) - Physics of Language Models: Part 1, Learning Hierarchical Language Structures [51.68385617116854]
Transformer-based language models are effective but complex, and understanding their inner workings is a significant challenge.
We introduce a family of synthetic CFGs that produce hierarchical rules, capable of generating lengthy sentences.
We demonstrate that generative models like GPT can accurately learn this CFG language and generate sentences based on it.
arXiv Detail & Related papers (2023-05-23T04:28:16Z) - Controlled Text Generation with Natural Language Instructions [74.88938055638636]
InstructCTG is a controlled text generation framework that incorporates different constraints.
We first extract the underlying constraints of natural texts through a combination of off-the-shelf NLP tools and simple verbalizes.
By prepending natural language descriptions of the constraints and a few demonstrations, we fine-tune a pre-trained language model to incorporate various types of constraints.
arXiv Detail & Related papers (2023-04-27T15:56:34Z) - Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language
Models [58.42146641102329]
We develop a novel semi-parametric language model architecture, Knowledge-in-Context (KiC)
KiC empowers a parametric text-to-text language model with a knowledge-rich external memory.
As a knowledge-rich semi-parametric language model, KiC only needs a much smaller part to achieve superior zero-shot performance on unseen tasks.
arXiv Detail & Related papers (2022-10-28T23:18:43Z) - Classifiers are Better Experts for Controllable Text Generation [63.17266060165098]
We show that the proposed method significantly outperforms recent PPLM, GeDi, and DExperts on PPL and sentiment accuracy based on the external classifier of generated texts.
The same time, it is also easier to implement and tune, and has significantly fewer restrictions and requirements.
arXiv Detail & Related papers (2022-05-15T12:58:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.