Collocation2Text: Controllable Text Generation from Guide Phrases in
Russian
- URL: http://arxiv.org/abs/2206.09248v1
- Date: Sat, 18 Jun 2022 17:10:08 GMT
- Title: Collocation2Text: Controllable Text Generation from Guide Phrases in
Russian
- Authors: Sergey Vychegzhanin, Evgeny Kotelnikov
- Abstract summary: Collocation2Text is a plug-and-play method for automatic controllable text generation in Russian.
The method is based on two interacting models: the autoregressive language ruGPT-3 model and the autoencoding language ruRoBERTa model.
Experiments on generating news articles using the proposed method showed its effectiveness for automatically generated fluent texts.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Large pre-trained language models are capable of generating varied and fluent
texts. Starting from the prompt, these models generate a narrative that can
develop unpredictably. The existing methods of controllable text generation,
which guide the narrative in the text in the user-specified direction, require
creating a training corpus and an additional time-consuming training procedure.
The paper proposes and investigates Collocation2Text, a plug-and-play method
for automatic controllable text generation in Russian, which does not require
fine-tuning. The method is based on two interacting models: the autoregressive
language ruGPT-3 model and the autoencoding language ruRoBERTa model. The idea
of the method is to shift the output distribution of the autoregressive model
according to the output distribution of the autoencoding model in order to
ensure a coherent transition of the narrative in the text towards the guide
phrase, which can contain single words or collocations. The autoencoding model,
which is able to take into account the left and right contexts of the token,
"tells" the autoregressive model which tokens are the most and least logical at
the current generation step, increasing or decreasing the probabilities of the
corresponding tokens. The experiments on generating news articles using the
proposed method showed its effectiveness for automatically generated fluent
texts which contain coherent transitions between user-specified phrases.
Related papers
- Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation [82.5217996570387]
We adapt a pre-trained language model for auto-regressive text-to-image generation.
We find that pre-trained language models offer limited help.
arXiv Detail & Related papers (2023-11-27T07:19:26Z) - Scalable Learning of Latent Language Structure With Logical Offline
Cycle Consistency [71.42261918225773]
Conceptually, LOCCO can be viewed as a form of self-learning where the semantic being trained is used to generate annotations for unlabeled text.
As an added bonus, the annotations produced by LOCCO can be trivially repurposed to train a neural text generation model.
arXiv Detail & Related papers (2023-05-31T16:47:20Z) - The Whole Truth and Nothing But the Truth: Faithful and Controllable
Dialogue Response Generation with Dataflow Transduction and Constrained
Decoding [65.34601470417967]
We describe a hybrid architecture for dialogue response generation that combines the strengths of neural language modeling and rule-based generation.
Our experiments show that this system outperforms both rule-based and learned approaches in human evaluations of fluency, relevance, and truthfulness.
arXiv Detail & Related papers (2022-09-16T09:00:49Z) - A Plug-and-Play Method for Controlled Text Generation [38.283313068622085]
We present a plug-and-play decoding method for controlled language generation that is so simple and intuitive, it can be described in a single sentence.
Despite the simplicity of this approach, we see it works incredibly well in practice.
arXiv Detail & Related papers (2021-09-20T17:27:03Z) - Inference Time Style Control for Summarization [6.017006996402699]
We present two novel methods that can be deployed during summary decoding on any pre-trained Transformer-based summarization model.
In experiments of summarizing with simplicity control, automatic evaluation and human judges both find our models producing outputs in simpler languages while still informative.
arXiv Detail & Related papers (2021-04-05T00:27:18Z) - Syntax-Enhanced Pre-trained Model [49.1659635460369]
We study the problem of leveraging the syntactic structure of text to enhance pre-trained models such as BERT and RoBERTa.
Existing methods utilize syntax of text either in the pre-training stage or in the fine-tuning stage, so that they suffer from discrepancy between the two stages.
We present a model that utilizes the syntax of text in both pre-training and fine-tuning stages.
arXiv Detail & Related papers (2020-12-28T06:48:04Z) - Facts2Story: Controlling Text Generation by Key Facts [0.0]
We propose a controlled generation task based on expanding a sequence of facts, expressed in natural language, into a longer narrative.
We show that while auto-regressive, unidirectional Language Models such as GPT2 produce better fluency, they struggle to adhere to the requested facts.
We propose a plan-and-cloze model (using fine-tuned XLNet) which produces competitive fluency while adhering to the requested content.
arXiv Detail & Related papers (2020-12-08T10:14:29Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Exemplar-Controllable Paraphrasing and Translation using Bitext [57.92051459102902]
We adapt models from prior work to be able to learn solely from bilingual text (bitext)
Our single proposed model can perform four tasks: controlled paraphrase generation in both languages and controlled machine translation in both language directions.
arXiv Detail & Related papers (2020-10-12T17:02:50Z) - Syntax-driven Iterative Expansion Language Models for Controllable Text
Generation [2.578242050187029]
We propose a new paradigm for introducing a syntactic inductive bias into neural text generation.
Our experiments show that this paradigm is effective at text generation, with quality between LSTMs and Transformers, and comparable diversity.
arXiv Detail & Related papers (2020-04-05T14:29:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.