Length-controllable Abstractive Summarization by Guiding with Summary
Prototype
- URL: http://arxiv.org/abs/2001.07331v1
- Date: Tue, 21 Jan 2020 04:01:58 GMT
- Title: Length-controllable Abstractive Summarization by Guiding with Summary
Prototype
- Authors: Itsumi Saito, Kyosuke Nishida, Kosuke Nishida, Atsushi Otsuka, Hisako
Asano, Junji Tomita, Hiroyuki Shindo, Yuji Matsumoto
- Abstract summary: We propose a new length-controllable abstractive summarization model.
Our model generates a summary in two steps.
Experiments with the CNN/Daily Mail dataset and the NEWSROOM dataset show that our model outperformed previous models in length-controlled settings.
- Score: 27.094797760775297
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a new length-controllable abstractive summarization model. Recent
state-of-the-art abstractive summarization models based on encoder-decoder
models generate only one summary per source text. However, controllable
summarization, especially of the length, is an important aspect for practical
applications. Previous studies on length-controllable abstractive summarization
incorporate length embeddings in the decoder module for controlling the summary
length. Although the length embeddings can control where to stop decoding, they
do not decide which information should be included in the summary within the
length constraint. Unlike the previous models, our length-controllable
abstractive summarization model incorporates a word-level extractive module in
the encoder-decoder model instead of length embeddings. Our model generates a
summary in two steps. First, our word-level extractor extracts a sequence of
important words (we call it the "prototype text") from the source text
according to the word-level importance scores and the length constraint.
Second, the prototype text is used as additional input to the encoder-decoder
model, which generates a summary by jointly encoding and copying words from
both the prototype text and source text. Since the prototype text is a guide to
both the content and length of the summary, our model can generate an
informative and length-controlled summary. Experiments with the CNN/Daily Mail
dataset and the NEWSROOM dataset show that our model outperformed previous
models in length-controlled settings.
Related papers
- A Novel LLM-based Two-stage Summarization Approach for Long Dialogues [9.835499880812646]
This study proposes a hierarchical framework that segments and condenses information from long documents.
The condensation stage utilizes an unsupervised generation model to generate condensed data.
The summarization stage fine-tunes the abstractive summarization model on the condensed data to generate the final results.
arXiv Detail & Related papers (2024-10-09T03:42:40Z) - Summarizing long regulatory documents with a multi-step pipeline [2.2591852560804675]
We show that the effectiveness of a two-step architecture for summarizing long regulatory texts varies depending on the model used.
For abstractive encoder-decoder models with short context lengths, the effectiveness of an extractive step varies, whereas for long-context encoder-decoder models, the extractive step worsens their performance.
arXiv Detail & Related papers (2024-08-19T08:07:25Z) - InstructCMP: Length Control in Sentence Compression through Instruction-based Large Language Models [27.26285945442178]
InstructCMP is an approach to the sentence compression task that can consider the length constraint through instructions.
We show that applying the length priming significantly improves performances of InstructCMP in both zero-shot and fine-tuning settings.
arXiv Detail & Related papers (2024-06-16T23:00:47Z) - Summarization with Precise Length Control [23.688834410051]
We present a framework to generate summaries with precisely the specified number of tokens or sentences.
We jointly train the models to predict the lengths, so our model can generate summaries with optimal length.
arXiv Detail & Related papers (2023-05-09T04:45:24Z) - MACSum: Controllable Summarization with Mixed Attributes [56.685735509260276]
MACSum is the first human-annotated summarization dataset for controlling mixed attributes.
We propose two simple and effective parameter-efficient approaches for the new task of mixed controllable summarization.
arXiv Detail & Related papers (2022-11-09T17:17:37Z) - Summarization Programs: Interpretable Abstractive Summarization with
Neural Modular Trees [89.60269205320431]
Current abstractive summarization models either suffer from a lack of clear interpretability or provide incomplete rationales.
We propose the Summarization Program (SP), an interpretable modular framework consisting of an (ordered) list of binary trees.
A Summarization Program contains one root node per summary sentence, and a distinct tree connects each summary sentence to the document sentences.
arXiv Detail & Related papers (2022-09-21T16:50:22Z) - Speech Summarization using Restricted Self-Attention [79.89680891246827]
We introduce a single model optimized end-to-end for speech summarization.
We demonstrate that the proposed model learns to directly summarize speech for the How-2 corpus of instructional videos.
arXiv Detail & Related papers (2021-10-12T18:21:23Z) - CTRLsum: Towards Generic Controllable Text Summarization [54.69190421411766]
We presentsum, a novel framework for controllable summarization.
Our approach enables users to control multiple aspects of generated summaries by interacting with the summarization system.
Using a single unified model,sum is able to achieve a broad scope of summary manipulation at inference time.
arXiv Detail & Related papers (2020-12-08T08:54:36Z) - SEAL: Segment-wise Extractive-Abstractive Long-form Text Summarization [39.85688193525843]
We study a sequence-to-sequence setting with input sequence lengths up to 100,000 tokens and output sequence lengths up to 768 tokens.
We propose SEAL, a Transformer-based model, featuring a new encoder-decoder attention that dynamically extracts/selects input snippets to sparsely attend to for each output segment.
The SEAL model achieves state-of-the-art results on existing long-form summarization tasks, and outperforms strong baseline models on a new dataset/task we introduce, Search2Wiki, with much longer input text.
arXiv Detail & Related papers (2020-06-18T00:13:21Z) - Exploring Explainable Selection to Control Abstractive Summarization [51.74889133688111]
We develop a novel framework that focuses on explainability.
A novel pair-wise matrix captures the sentence interactions, centrality, and attribute scores.
A sentence-deployed attention mechanism in the abstractor ensures the final summary emphasizes the desired content.
arXiv Detail & Related papers (2020-04-24T14:39:34Z) - Pre-training for Abstractive Document Summarization by Reinstating
Source Text [105.77348528847337]
This paper presents three pre-training objectives which allow us to pre-train a Seq2Seq based abstractive summarization model on unlabeled text.
Experiments on two benchmark summarization datasets show that all three objectives can improve performance upon baselines.
arXiv Detail & Related papers (2020-04-04T05:06:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.