Constrained Abstractive Summarization: Preserving Factual Consistency
with Constrained Generation
- URL: http://arxiv.org/abs/2010.12723v2
- Date: Thu, 16 Dec 2021 05:20:15 GMT
- Title: Constrained Abstractive Summarization: Preserving Factual Consistency
with Constrained Generation
- Authors: Yuning Mao, Xiang Ren, Heng Ji, Jiawei Han
- Abstract summary: We propose Constrained Abstractive Summarization (CAS), a general setup that preserves the factual consistency of abstractive summarization.
We adopt lexically constrained decoding, a technique generally applicable to autoregressive generative models, to fulfill CAS.
We observe up to 13.8 ROUGE-2 gains when only one manual constraint is used in interactive summarization.
- Score: 93.87095877617968
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite significant progress, state-of-the-art abstractive summarization
methods are still prone to hallucinate content inconsistent with the source
document. In this paper, we propose Constrained Abstractive Summarization
(CAS), a general setup that preserves the factual consistency of abstractive
summarization by specifying tokens as constraints that must be present in the
summary. We adopt lexically constrained decoding, a technique generally
applicable to autoregressive generative models, to fulfill CAS and conduct
experiments in two scenarios: (1) automatic summarization without human
involvement, where keyphrases are extracted from the source document and used
as constraints; (2) human-guided interactive summarization, where human
feedback in the form of manual constraints are used to guide summary
generation. Automatic and human evaluations on two benchmark datasets
demonstrate that CAS improves both lexical overlap (ROUGE) and factual
consistency of abstractive summarization. In particular, we observe up to 13.8
ROUGE-2 gains when only one manual constraint is used in interactive
summarization.
Related papers
- AMRFact: Enhancing Summarization Factuality Evaluation with AMR-Driven Negative Samples Generation [57.8363998797433]
We propose AMRFact, a framework that generates perturbed summaries using Abstract Meaning Representations (AMRs)
Our approach parses factually consistent summaries into AMR graphs and injects controlled factual inconsistencies to create negative examples, allowing for coherent factually inconsistent summaries to be generated with high error-type coverage.
arXiv Detail & Related papers (2023-11-16T02:56:29Z) - How Ready are Pre-trained Abstractive Models and LLMs for Legal Case
Judgement Summarization? [4.721618284417204]
In recent years, abstractive summarization models are gaining popularity.
Legal domain-specific pre-trained abstractive summarization models are now available.
General-domain pre-trained Large Language Models (LLMs) are known to generate high-quality text.
arXiv Detail & Related papers (2023-06-02T03:16:19Z) - Salience Allocation as Guidance for Abstractive Summarization [61.31826412150143]
We propose a novel summarization approach with a flexible and reliable salience guidance, namely SEASON (SaliencE Allocation as Guidance for Abstractive SummarizatiON)
SEASON utilizes the allocation of salience expectation to guide abstractive summarization and adapts well to articles in different abstractiveness.
arXiv Detail & Related papers (2022-10-22T02:13:44Z) - Controllable Summarization with Constrained Markov Decision Process [50.04321779376415]
We study controllable text summarization which allows users to gain control on a particular attribute.
We propose a novel training framework based on Constrained Markov Decision Process (CMDP)
Our framework can be applied to control important attributes of summarization, including length, covered entities, and abstractiveness.
arXiv Detail & Related papers (2021-08-07T09:12:53Z) - Evaluating the Tradeoff Between Abstractiveness and Factuality in
Abstractive Summarization [20.83986393847262]
We analyze the tradeoff between abstractiveness and factuality of generated summaries across multiple datasets and models.
We propose new factuality metrics that adjust for the degree of abstractiveness.
arXiv Detail & Related papers (2021-08-05T21:28:20Z) - EASE: Extractive-Abstractive Summarization with Explanations [18.046254486733186]
We present an explainable summarization system based on the Information Bottleneck principle.
Inspired by previous research that humans use a two-stage framework to summarize long documents, our framework first extracts a pre-defined amount of evidence spans as explanations.
We show that explanations from our framework are more relevant than simple baselines, without substantially sacrificing the quality of the generated summary.
arXiv Detail & Related papers (2021-05-14T17:45:06Z) - The Summary Loop: Learning to Write Abstractive Summaries Without
Examples [21.85348918324668]
This work presents a new approach to unsupervised abstractive summarization based on maximizing a combination of coverage and fluency for a given length constraint.
Key terms are masked out of the original document and must be filled in by a coverage model using the current generated summary.
When tested on popular news summarization datasets, the method outperforms previous unsupervised methods by more than 2 R-1 points.
arXiv Detail & Related papers (2021-05-11T23:19:46Z) - Multi-Fact Correction in Abstractive Text Summarization [98.27031108197944]
Span-Fact is a suite of two factual correction models that leverages knowledge learned from question answering models to make corrections in system-generated summaries via span selection.
Our models employ single or multi-masking strategies to either iteratively or auto-regressively replace entities in order to ensure semantic consistency w.r.t. the source text.
Experiments show that our models significantly boost the factual consistency of system-generated summaries without sacrificing summary quality in terms of both automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-10-06T02:51:02Z) - Enhancing Factual Consistency of Abstractive Summarization [57.67609672082137]
We propose a fact-aware summarization model FASum to extract and integrate factual relations into the summary generation process.
We then design a factual corrector model FC to automatically correct factual errors from summaries generated by existing systems.
arXiv Detail & Related papers (2020-03-19T07:36:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.