KEST: Kernel Distance Based Efficient Self-Training for Improving
Controllable Text Generation
- URL: http://arxiv.org/abs/2306.10414v1
- Date: Sat, 17 Jun 2023 19:40:57 GMT
- Title: KEST: Kernel Distance Based Efficient Self-Training for Improving
Controllable Text Generation
- Authors: Yuxi Feng, Xiaoyuan Yi, Laks V.S. Lakshmanan, and Xing Xie
- Abstract summary: We propose KEST, a novel and efficient self-training framework to handle these problems.
KEST utilizes a kernel-based loss, rather than standard cross entropy, to learn from the soft pseudo text produced by a shared non-autoregressive generator.
Experiments on three controllable generation tasks demonstrate that KEST significantly improves control accuracy while maintaining comparable text fluency and generation diversity against several strong baselines.
- Score: 24.47531522553703
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-training (ST) has come to fruition in language understanding tasks by
producing pseudo labels, which reduces the labeling bottleneck of language
model fine-tuning. Nevertheless, in facilitating semi-supervised controllable
language generation, ST faces two key challenges. First, augmented by
self-generated pseudo text, generation models tend to over-exploit the
previously learned text distribution, suffering from mode collapse and poor
generation diversity. Second, generating pseudo text in each iteration is
time-consuming, severely decelerating the training process. In this work, we
propose KEST, a novel and efficient self-training framework to handle these
problems. KEST utilizes a kernel-based loss, rather than standard cross
entropy, to learn from the soft pseudo text produced by a shared
non-autoregressive generator. We demonstrate both theoretically and empirically
that KEST can benefit from more diverse pseudo text in an efficient manner,
which allows not only refining and exploiting the previously fitted
distribution but also enhanced exploration towards a larger potential text
space, providing a guarantee of improved performance. Experiments on three
controllable generation tasks demonstrate that KEST significantly improves
control accuracy while maintaining comparable text fluency and generation
diversity against several strong baselines.
Related papers
- Detecting Machine-Generated Long-Form Content with Latent-Space Variables [54.07946647012579]
Existing zero-shot detectors primarily focus on token-level distributions, which are vulnerable to real-world domain shifts.
We propose a more robust method that incorporates abstract elements, such as event transitions, as key deciding factors to detect machine versus human texts.
arXiv Detail & Related papers (2024-10-04T18:42:09Z) - Retrieval is Accurate Generation [99.24267226311157]
We introduce a novel method that selects context-aware phrases from a collection of supporting documents.
Our model achieves the best performance and the lowest latency among several retrieval-augmented baselines.
arXiv Detail & Related papers (2024-02-27T14:16:19Z) - Successor Features for Efficient Multisubject Controlled Text Generation [48.37713738712319]
We introduce SF-GEN, which is grounded in two primary concepts: successor features (SFs) and language model rectification.
SF-GEN seamlessly integrates the two to enable dynamic steering of text generation with no need to alter the LLM's parameters.
To the best of our knowledge, our research represents the first application of successor features in text generation.
arXiv Detail & Related papers (2023-11-03T00:17:08Z) - STA: Self-controlled Text Augmentation for Improving Text
Classifications [2.9669250132689164]
A number of text augmentation techniques have emerged in the field of Natural Language Processing (NLP)
We introduce a state-of-the-art approach for Self-Controlled Text Augmentation (STA)
Our approach tightly controls the generation process by introducing a self-checking procedure to ensure that generated examples retain the semantic content of the original text.
arXiv Detail & Related papers (2023-02-24T17:54:12Z) - DuNST: Dual Noisy Self Training for Semi-Supervised Controllable Text
Generation [34.49194429157166]
Self-training (ST) has prospered again in language understanding by augmenting the fine-tuning of pre-trained language models when labeled data is insufficient.
It remains challenging to incorporate ST into attribute-controllable language generation.
arXiv Detail & Related papers (2022-12-16T21:44:34Z) - FAST: Improving Controllability for Text Generation with Feedback Aware
Self-Training [25.75982440355576]
Controllable text generation systems often leverage control codes to direct various properties of the output like style and length.
Inspired by recent work on causal inference for NLP, this paper reveals a previously overlooked flaw in these control code-based conditional text generation algorithms.
We propose two simple techniques to reduce these correlations in training sets.
arXiv Detail & Related papers (2022-10-06T19:00:51Z) - Text Generation with Efficient (Soft) Q-Learning [91.47743595382758]
Reinforcement learning (RL) offers a more flexible solution by allowing users to plug in arbitrary task metrics as reward.
We introduce a new RL formulation for text generation from the soft Q-learning perspective.
We apply the approach to a wide range of tasks, including learning from noisy/negative examples, adversarial attacks, and prompt generation.
arXiv Detail & Related papers (2021-06-14T18:48:40Z) - Improving Text Generation with Student-Forcing Optimal Transport [122.11881937642401]
We propose using optimal transport (OT) to match the sequences generated in training and testing modes.
An extension is also proposed to improve the OT learning, based on the structural and contextual information of the text sequences.
The effectiveness of the proposed method is validated on machine translation, text summarization, and text generation tasks.
arXiv Detail & Related papers (2020-10-12T19:42:25Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.