An Extensible Plug-and-Play Method for Multi-Aspect Controllable Text
Generation
- URL: http://arxiv.org/abs/2212.09387v2
- Date: Sun, 28 May 2023 14:12:48 GMT
- Title: An Extensible Plug-and-Play Method for Multi-Aspect Controllable Text
Generation
- Authors: Xuancheng Huang, Zijun Liu, Peng Li, Tao Li, Maosong Sun, Yang Liu
- Abstract summary: Multi-aspect controllable text generation that controls generated text in multiple aspects has attracted increasing attention.
We provide a theoretical lower bound for the interference and empirically found that the interference grows with the number of layers where prefixes are inserted.
We propose using trainable gates to normalize the intervention of prefixes to restrain the growing interference.
- Score: 70.77243918587321
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, multi-aspect controllable text generation that controls the
generated text in multiple aspects (e.g., sentiment, topic, and keywords) has
attracted increasing attention. Although methods based on parameter efficient
tuning like prefix-tuning could achieve multi-aspect controlling in a
plug-and-play way, the mutual interference of multiple prefixes leads to
significant degeneration of constraints and limits their extensibility to
training-time unseen aspect combinations. In this work, we provide a
theoretical lower bound for the interference and empirically found that the
interference grows with the number of layers where prefixes are inserted. Based
on these analyses, we propose using trainable gates to normalize the
intervention of prefixes to restrain the growing interference. As a result,
controlling training-time unseen combinations of aspects can be realized by
simply concatenating corresponding plugins such that new constraints can be
extended at a lower cost. In addition, we propose a unified way to process both
categorical and free-form constraints. Experiments on text generation and
machine translation demonstrate the superiority of our approach over baselines
on constraint accuracy, text quality, and extensibility.
Related papers
- Intertwining CP and NLP: The Generation of Unreasonably Constrained Sentences [49.86129209397701]
This paper presents the Constraints First Framework to remedy this issue.
It is solved by a constraint programming method that combines linguistic properties with more classical constraints.
The effectiveness of this approach is demonstrated by tackling a new more tediously constrained text generation problem.
arXiv Detail & Related papers (2024-06-15T17:40:49Z) - Toward Unified Controllable Text Generation via Regular Expression
Instruction [56.68753672187368]
Our paper introduces Regular Expression Instruction (REI), which utilizes an instruction-based mechanism to fully exploit regular expressions' advantages to uniformly model diverse constraints.
Our method only requires fine-tuning on medium-scale language models or few-shot, in-context learning on large language models, and requires no further adjustment when applied to various constraint combinations.
arXiv Detail & Related papers (2023-09-19T09:05:14Z) - MacLaSa: Multi-Aspect Controllable Text Generation via Efficient
Sampling from Compact Latent Space [110.85888003111653]
Multi-aspect controllable text generation aims to generate fluent sentences that possess multiple desired attributes simultaneously.
We introduce a novel approach for multi-aspect control, namely MacLaSa, that estimates compact latent space for multiple aspects.
We show that MacLaSa outperforms several strong baselines on attribute relevance and textual quality while maintaining a high inference speed.
arXiv Detail & Related papers (2023-05-22T07:30:35Z) - A Distributional Lens for Multi-Aspect Controllable Text Generation [17.97374410245602]
Multi-aspect controllable text generation is a more challenging and practical task than single-aspect control.
Existing methods achieve complex multi-aspect control by fusing multiple controllers learned from single-aspect.
We propose to directly search for the intersection areas of multiple attribute distributions as their combination for generation.
arXiv Detail & Related papers (2022-10-06T13:08:04Z) - Controllable Natural Language Generation with Contrastive Prefixes [120.12778570283956]
GPT2 generation utilizes a set of small attribute-specific vectors, called prefixes, to steer natural language generation.
We propose a novel supervised method and also an unsupervised method to train the prefixes for single-aspect control.
Experimental results on both single-aspect and multi-aspect control show that our methods can guide generation towards the desired attributes while keeping high linguistic quality.
arXiv Detail & Related papers (2022-02-27T00:31:03Z) - COLD Decoding: Energy-based Constrained Text Generation with Langevin
Dynamics [69.8062252611486]
Cold decoding is a flexible framework that can be applied directly to off-the-shelf left-to-right language models.
Our experiments on constrained generation tasks point to the effectiveness of our approach, both in terms of automatic and human evaluation.
arXiv Detail & Related papers (2022-02-23T18:59:27Z) - Controlled Text Generation as Continuous Optimization with Multiple
Constraints [23.71027518888138]
We propose a flexible and modular algorithm for controllable inference from pretrained models.
We make use of Lagrangian multipliers and gradient-descent based techniques to generate the desired text.
We evaluate our approach on controllable machine translation and style transfer with multiple sentence-level attributes.
arXiv Detail & Related papers (2021-08-04T05:25:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.