Latent Prompt Tuning for Text Summarization
- URL: http://arxiv.org/abs/2211.01837v1
- Date: Thu, 3 Nov 2022 14:18:48 GMT
- Title: Latent Prompt Tuning for Text Summarization
- Authors: Yubo Zhang, Xingxing Zhang, Xun Wang, Si-qing Chen and Furu Wei
- Abstract summary: We propose Lotus (shorthand for Latent Prompt Tuning for Summarization), which is a single model that can be applied in both controlled and uncontrolled modes.
During training, Lotus learns latent prompt representations from prompts with gold control signals using a contrastive learning objective.
Experiments show Lotus in uncontrolled mode consistently improves upon strong (uncontrollable) summarization models across four different summarization datasets.
- Score: 95.85520030785139
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Prompts with different control signals (e.g., length, keywords, etc.) can be
used to control text summarization. When control signals are available, they
can control the properties of generated summaries and potentially improve
summarization quality (since more information are given). Unfortunately,
control signals are not already available during inference time. In this paper,
we propose Lotus (shorthand for Latent Prompt Tuning for Summarization), which
is a single model that can be applied in both controlled and uncontrolled
(without control signals) modes. During training, Lotus learns latent prompt
representations from prompts with gold control signals using a contrastive
learning objective. Experiments show Lotus in uncontrolled mode consistently
improves upon strong (uncontrollable) summarization models across four
different summarization datasets. We also demonstrate generated summaries can
be controlled using prompts with user specified control tokens.
Related papers
- AnyControl: Create Your Artwork with Versatile Control on Text-to-Image Generation [24.07613591217345]
Linguistic control enables effective content creation, but struggles with fine-grained control over image generation.
AnyControl develops a novel Multi-Control framework that extracts a unified multi-modal embedding to guide the generation process.
This approach enables a holistic understanding of user inputs, and produces high-quality, faithful results under versatile control signals.
arXiv Detail & Related papers (2024-06-27T07:40:59Z) - Prompt-Based Length Controlled Generation with Multiple Control Types [45.202705040391734]
We propose a prompt-based method to achieve length controlled generation under different control types with high accuracy.
In particular, we adopt reinforcement learning (RL) and sample filtering with the reward signal given by rule-based reward models.
Experiments show that our method significantly improves the accuracy of prompt-based length control on popular summarization datasets like CNNDM and NYT.
arXiv Detail & Related papers (2024-06-12T01:49:54Z) - ControlNet-XS: Rethinking the Control of Text-to-Image Diffusion Models as Feedback-Control Systems [19.02295657801464]
In this work, we take an existing controlling network (ControlNet) and change the communication between the controlling network and the generation process to be of high-frequency and with large-bandwidth.
We outperform state-of-the-art approaches for pixel-level guidance, such as depth, canny-edges, and semantic segmentation, and are on a par for loose keypoint-guidance of human poses.
All code and pre-trained models will be made publicly available.
arXiv Detail & Related papers (2023-12-11T17:58:06Z) - Fine-grained Controllable Video Generation via Object Appearance and
Context [74.23066823064575]
We propose fine-grained controllable video generation (FACTOR) to achieve detailed control.
FACTOR aims to control objects' appearances and context, including their location and category.
Our method achieves controllability of object appearances without finetuning, which reduces the per-subject optimization efforts for the users.
arXiv Detail & Related papers (2023-12-05T17:47:33Z) - Readout Guidance: Learning Control from Diffusion Features [96.22155562120231]
We present Readout Guidance, a method for controlling text-to-image diffusion models with learned signals.
Readout Guidance uses readout heads, lightweight networks trained to extract signals from the features of a pre-trained, frozen diffusion model at every timestep.
These readouts can encode single-image properties, such as pose, depth, and edges; or higher-order properties that relate multiple images, such as correspondence and appearance similarity.
arXiv Detail & Related papers (2023-12-04T18:59:32Z) - Cocktail: Mixing Multi-Modality Controls for Text-Conditional Image
Generation [79.8881514424969]
Text-conditional diffusion models are able to generate high-fidelity images with diverse contents.
However, linguistic representations frequently exhibit ambiguous descriptions of the envisioned objective imagery.
We propose Cocktail, a pipeline to mix various modalities into one embedding.
arXiv Detail & Related papers (2023-06-01T17:55:32Z) - CTRLsum: Towards Generic Controllable Text Summarization [54.69190421411766]
We presentsum, a novel framework for controllable summarization.
Our approach enables users to control multiple aspects of generated summaries by interacting with the summarization system.
Using a single unified model,sum is able to achieve a broad scope of summary manipulation at inference time.
arXiv Detail & Related papers (2020-12-08T08:54:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.