Focused Prefix Tuning for Controllable Text Generation
- URL: http://arxiv.org/abs/2306.00369v2
- Date: Sat, 10 Jun 2023 12:36:48 GMT
- Title: Focused Prefix Tuning for Controllable Text Generation
- Authors: Congda Ma, Tianyu Zhao, Makoto Shing, Kei Sawada, Manabu Okumura
- Abstract summary: We propose focused prefix tuning(FPT) to mitigate the problem and to enable the control to focus on the desired attribute.
Experimental results show that FPT can achieve better control accuracy and text fluency than baseline models in single-attribute control tasks.
- Score: 19.88484696133778
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In a controllable text generation dataset, there exist unannotated attributes
that could provide irrelevant learning signals to models that use it for
training and thus degrade their performance. We propose focused prefix
tuning(FPT) to mitigate the problem and to enable the control to focus on the
desired attribute. Experimental results show that FPT can achieve better
control accuracy and text fluency than baseline models in single-attribute
control tasks. In multi-attribute control tasks, FPT achieves comparable
control accuracy with the state-of-the-art approach while keeping the
flexibility to control new attributes without retraining existing models.
Related papers
- Adding Conditional Control to Diffusion Models with Reinforcement Learning [59.295203871547336]
Diffusion models are powerful generative models that allow for precise control over the characteristics of the generated samples.
This work presents a novel method based on reinforcement learning (RL) to add additional controls, leveraging an offline dataset.
arXiv Detail & Related papers (2024-06-17T22:00:26Z) - LiFi: Lightweight Controlled Text Generation with Fine-Grained Control
Codes [46.74968005604948]
We present LIFI, which offers a lightweight approach with fine-grained control for controlled text generation.
We evaluate LIFI on two conventional tasks -- sentiment control and topic control -- and one newly proposed task -- stylistic novel writing.
arXiv Detail & Related papers (2024-02-10T11:53:48Z) - Fine-grained Controllable Video Generation via Object Appearance and
Context [74.23066823064575]
We propose fine-grained controllable video generation (FACTOR) to achieve detailed control.
FACTOR aims to control objects' appearances and context, including their location and category.
Our method achieves controllability of object appearances without finetuning, which reduces the per-subject optimization efforts for the users.
arXiv Detail & Related papers (2023-12-05T17:47:33Z) - Parameter-Efficient Tuning Helps Language Model Alignment [57.27390187540737]
Previous works mainly adopt reinforcement learning (RLHF) and direct preference optimization (DPO) with human feedback for alignment.
Controllable generation offers more flexibility with regard to data format.
Our approach, alignMEnt with parameter-Efficient Tuning (MEET), improves the quality of control tokens.
arXiv Detail & Related papers (2023-10-01T23:27:14Z) - Active Finetuning: Exploiting Annotation Budget in the
Pretraining-Finetuning Paradigm [132.9949120482274]
This paper focuses on the selection of samples for annotation in the pretraining-finetuning paradigm.
We propose a novel method called ActiveFT for active finetuning task to select a subset of data distributing similarly with the entire unlabeled pool.
Extensive experiments show the leading performance and high efficiency of ActiveFT superior to baselines on both image classification and semantic segmentation.
arXiv Detail & Related papers (2023-03-25T07:17:03Z) - FAST: Improving Controllability for Text Generation with Feedback Aware
Self-Training [25.75982440355576]
Controllable text generation systems often leverage control codes to direct various properties of the output like style and length.
Inspired by recent work on causal inference for NLP, this paper reveals a previously overlooked flaw in these control code-based conditional text generation algorithms.
We propose two simple techniques to reduce these correlations in training sets.
arXiv Detail & Related papers (2022-10-06T19:00:51Z) - Plug and Play Counterfactual Text Generation for Model Robustness [12.517365153658028]
We introduce CASPer, a plug-and-play counterfactual generation framework.
We show that CASPer effectively generates counterfactual text that follow the steering provided by an attribute model.
We also show that the generated counterfactuals can be used for augmenting the training data and thereby fixing and making the test model more robust.
arXiv Detail & Related papers (2022-06-21T14:25:21Z) - Mix and Match: Learning-free Controllable Text Generation using Energy
Language Models [33.97800741890231]
We propose Mix and Match LM, a global score-based alternative for controllable text generation.
We interpret the task of controllable generation as drawing samples from an energy-based model.
We use a Metropolis-Hastings sampling scheme to sample from this energy-based model.
arXiv Detail & Related papers (2022-03-24T18:52:09Z) - Controlled Text Generation as Continuous Optimization with Multiple
Constraints [23.71027518888138]
We propose a flexible and modular algorithm for controllable inference from pretrained models.
We make use of Lagrangian multipliers and gradient-descent based techniques to generate the desired text.
We evaluate our approach on controllable machine translation and style transfer with multiple sentence-level attributes.
arXiv Detail & Related papers (2021-08-04T05:25:20Z) - How Training Data Impacts Performance in Learning-based Control [67.7875109298865]
This paper derives an analytical relationship between the density of the training data and the control performance.
We formulate a quality measure for the data set, which we refer to as $rho$-gap.
We show how the $rho$-gap can be applied to a feedback linearizing control law.
arXiv Detail & Related papers (2020-05-25T12:13:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.