Enhancing Coherence of Extractive Summarization with Multitask Learning
- URL: http://arxiv.org/abs/2305.12851v2
- Date: Fri, 21 Jul 2023 10:22:53 GMT
- Title: Enhancing Coherence of Extractive Summarization with Multitask Learning
- Authors: Renlong Jie, Xiaojun Meng, Lifeng Shang, Xin Jiang, Qun Liu
- Abstract summary: This study proposes a multitask learning architecture for extractive summarization with coherence boosting.
The architecture contains an extractive summarizer and coherent discriminator module.
Experiments show that our proposed method significantly improves the proportion of consecutive sentences in the extracted summaries.
- Score: 40.349019691412465
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This study proposes a multitask learning architecture for extractive
summarization with coherence boosting. The architecture contains an extractive
summarizer and coherent discriminator module. The coherent discriminator is
trained online on the sentence vectors of the augmented textual input, thus
improving its general ability of judging whether the input sentences are
coherent. Meanwhile, we maximize the coherent scores from the coherent
discriminator by updating the parameters of the summarizer. To make the
extractive sentences trainable in a differentiable manner, we introduce two
strategies, including pre-trained converting model (model-based) and converting
matrix (MAT-based) that merge sentence representations. Experiments show that
our proposed method significantly improves the proportion of consecutive
sentences in the extracted summaries based on their positions in the original
article (i.e., automatic sentence-level coherence metric), while the goodness
in terms of other automatic metrics (i.e., Rouge scores and BertScores) are
preserved. Human evaluation also evidences the improvement of coherence and
consistency of the extracted summaries given by our method.
Related papers
- Hierarchical Indexing for Retrieval-Augmented Opinion Summarization [60.5923941324953]
We propose a method for unsupervised abstractive opinion summarization that combines the attributability and scalability of extractive approaches with the coherence and fluency of Large Language Models (LLMs)
Our method, HIRO, learns an index structure that maps sentences to a path through a semantically organized discrete hierarchy.
At inference time, we populate the index and use it to identify and retrieve clusters of sentences containing popular opinions from input reviews.
arXiv Detail & Related papers (2024-03-01T10:38:07Z) - Efficient and Interpretable Compressive Text Summarisation with
Unsupervised Dual-Agent Reinforcement Learning [36.93582300019002]
We propose an efficient and interpretable compressive summarisation method using unsupervised dual-agent reinforcement learning.
Our model achieves promising performance and a significant improvement on Newsroom in terms of the ROUGE metric.
arXiv Detail & Related papers (2023-06-06T05:30:49Z) - Attributable and Scalable Opinion Summarization [79.87892048285819]
We generate abstractive summaries by decoding frequent encodings, and extractive summaries by selecting the sentences assigned to the same frequent encodings.
Our method is attributable, because the model identifies sentences used to generate the summary as part of the summarization process.
It scales easily to many hundreds of input reviews, because aggregation is performed in the latent space rather than over long sequences of tokens.
arXiv Detail & Related papers (2023-05-19T11:30:37Z) - A General Contextualized Rewriting Framework for Text Summarization [15.311467109946571]
Exiting rewriting systems take each extractive sentence as the only input, which is relatively focused but can lose necessary background knowledge and discourse context.
We formalize contextualized rewriting as a seq2seq with group-tag alignments, identifying extractive sentences through content-based addressing.
Results show that our approach significantly outperforms non-contextualized rewriting systems without requiring reinforcement learning.
arXiv Detail & Related papers (2022-07-13T03:55:57Z) - Automated Concatenation of Embeddings for Structured Prediction [75.44925576268052]
We propose Automated Concatenation of Embeddings (ACE) to automate the process of finding better concatenations of embeddings for structured prediction tasks.
We follow strategies in reinforcement learning to optimize the parameters of the controller and compute the reward based on the accuracy of a task model.
arXiv Detail & Related papers (2020-10-10T14:03:20Z) - Multi-Fact Correction in Abstractive Text Summarization [98.27031108197944]
Span-Fact is a suite of two factual correction models that leverages knowledge learned from question answering models to make corrections in system-generated summaries via span selection.
Our models employ single or multi-masking strategies to either iteratively or auto-regressively replace entities in order to ensure semantic consistency w.r.t. the source text.
Experiments show that our models significantly boost the factual consistency of system-generated summaries without sacrificing summary quality in terms of both automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-10-06T02:51:02Z) - Unsupervised Summarization by Jointly Extracting Sentences and Keywords [12.387378783627762]
RepRank is an unsupervised graph-based ranking model for extractive multi-document summarization.
We show that salient sentences and keywords can be extracted in a joint and mutual reinforcement process using our learned representations.
Experiment results with multiple benchmark datasets show that RepRank achieved the best or comparable performance in ROUGE.
arXiv Detail & Related papers (2020-09-16T05:58:00Z) - Extractive Summarization as Text Matching [123.09816729675838]
This paper creates a paradigm shift with regard to the way we build neural extractive summarization systems.
We formulate the extractive summarization task as a semantic text matching problem.
We have driven the state-of-the-art extractive result on CNN/DailyMail to a new level (44.41 in ROUGE-1)
arXiv Detail & Related papers (2020-04-19T08:27:57Z) - AREDSUM: Adaptive Redundancy-Aware Iterative Sentence Ranking for
Extractive Document Summarization [46.00136909474304]
Redundancy-aware extractive summarization systems score the redundancy of the sentences to be included in a summary.
Previous work shows the efficacy of jointly scoring and selecting sentences with neural sequence generation models.
We present two adaptive learning models: AREDSUM-SEQ that jointly considers salience and novelty during sentence selection; and a two-step AREDSUM-CTX that scores salience first, then learns to balance salience and redundancy.
arXiv Detail & Related papers (2020-04-13T20:02:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.