MacLaSa: Multi-Aspect Controllable Text Generation via Efficient
Sampling from Compact Latent Space
- URL: http://arxiv.org/abs/2305.12785v2
- Date: Tue, 17 Oct 2023 15:48:15 GMT
- Title: MacLaSa: Multi-Aspect Controllable Text Generation via Efficient
Sampling from Compact Latent Space
- Authors: Hanxing Ding, Liang Pang, Zihao Wei, Huawei Shen, Xueqi Cheng,
Tat-Seng Chua
- Abstract summary: Multi-aspect controllable text generation aims to generate fluent sentences that possess multiple desired attributes simultaneously.
We introduce a novel approach for multi-aspect control, namely MacLaSa, that estimates compact latent space for multiple aspects.
We show that MacLaSa outperforms several strong baselines on attribute relevance and textual quality while maintaining a high inference speed.
- Score: 110.85888003111653
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-aspect controllable text generation aims to generate fluent sentences
that possess multiple desired attributes simultaneously. Traditional methods
either combine many operators in the decoding stage, often with costly
iteration or search in the discrete text space, or train separate controllers
for each aspect, resulting in a degeneration of text quality due to the
discrepancy between different aspects. To address these limitations, we
introduce a novel approach for multi-aspect control, namely MacLaSa, that
estimates compact latent space for multiple aspects and performs efficient
sampling with a robust sampler based on ordinary differential equations (ODEs).
To eliminate the domain gaps between different aspects, we utilize a
Variational Autoencoder (VAE) network to map text sequences from varying data
sources into close latent representations. The estimated latent space enables
the formulation of joint energy-based models (EBMs) and the plugging in of
arbitrary attribute discriminators to achieve multi-aspect control. Afterwards,
we draw latent vector samples with an ODE-based sampler and feed sampled
examples to the VAE decoder to produce target text sequences. Experimental
results demonstrate that MacLaSa outperforms several strong baselines on
attribute relevance and textual quality while maintaining a high inference
speed.
Related papers
- Detecting Machine-Generated Long-Form Content with Latent-Space Variables [54.07946647012579]
Existing zero-shot detectors primarily focus on token-level distributions, which are vulnerable to real-world domain shifts.
We propose a more robust method that incorporates abstract elements, such as event transitions, as key deciding factors to detect machine versus human texts.
arXiv Detail & Related papers (2024-10-04T18:42:09Z) - Principled Gradient-based Markov Chain Monte Carlo for Text Generation [77.46654898866291]
We propose several faithful gradient-based sampling algorithms to sample from the target energy-based text distribution correctly.
We demonstrate that faithful samplers are able to generate more fluent text while adhering to the control objectives better.
arXiv Detail & Related papers (2023-12-29T18:00:56Z) - Semantic Space Grounded Weighted Decoding for Multi-Attribute
Controllable Dialogue Generation [41.23970507903113]
We propose a novel framework called DASC that possesses strong controllability with a weighted decoding paradigm.
Generation with multiple attributes is then intuitively implemented with an utterance of multiple attribute embeddings.
Experiments show that DASC can achieve high control accuracy in generation task with the simultaneous control of 3 aspects.
arXiv Detail & Related papers (2023-05-04T13:35:27Z) - Controllable Text Generation via Probability Density Estimation in the
Latent Space [16.962510129437558]
We propose a novel control framework using probability density estimation in the latent space.
Our method utilizes an invertible transformation function, the Normalizing Flow, that maps the complex distributions in the latent space to simple Gaussian distributions in the prior space.
Experiments on single-attribute controls and multi-attribute control reveal that our method outperforms several strong baselines on attribute relevance and text quality.
arXiv Detail & Related papers (2022-12-16T07:11:18Z) - Arithmetic Sampling: Parallel Diverse Decoding for Large Language Models [65.52639709094963]
Methods such as beam search and Gumbel top-k sampling can guarantee a different output for each element of the beam, but are not easy to parallelize.
We present a framework for sampling according to an arithmetic code book implicitly defined by a large language model.
arXiv Detail & Related papers (2022-10-18T22:19:41Z) - A Distributional Lens for Multi-Aspect Controllable Text Generation [17.97374410245602]
Multi-aspect controllable text generation is a more challenging and practical task than single-aspect control.
Existing methods achieve complex multi-aspect control by fusing multiple controllers learned from single-aspect.
We propose to directly search for the intersection areas of multiple attribute distributions as their combination for generation.
arXiv Detail & Related papers (2022-10-06T13:08:04Z) - Composable Text Controls in Latent Space with ODEs [97.12426987887021]
This paper proposes a new efficient approach for composable text operations in the compact latent space of text.
By connecting pretrained LMs to the latent space through efficient adaption, we then decode the sampled vectors into desired text sequences.
Experiments show that composing those operators within our approach manages to generate or edit high-quality text.
arXiv Detail & Related papers (2022-08-01T06:51:45Z) - Multi-scale and Cross-scale Contrastive Learning for Semantic
Segmentation [5.281694565226513]
We apply contrastive learning to enhance the discriminative power of the multi-scale features extracted by semantic segmentation networks.
By first mapping the encoder's multi-scale representations to a common feature space, we instantiate a novel form of supervised local-global constraint.
arXiv Detail & Related papers (2022-03-25T01:24:24Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.