Reinforcing Semantic-Symmetry for Document Summarization
- URL: http://arxiv.org/abs/2112.07583v1
- Date: Tue, 14 Dec 2021 17:41:37 GMT
- Title: Reinforcing Semantic-Symmetry for Document Summarization
- Authors: Mingyang Song, Liping Jing
- Abstract summary: Document summarization condenses a long document into a short version with salient information and accurate semantic descriptions.
This paper introduces a new textbfreinforcing stextbfemantic-textbfsymmetry learning textbfmodel is proposed for document summarization.
A series of experiments have been conducted on two wildly used benchmark datasets CNN/Daily Mail and BigPatent.
- Score: 15.113768658584979
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Document summarization condenses a long document into a short version with
salient information and accurate semantic descriptions. The main issue is how
to make the output summary semantically consistent with the input document. To
reach this goal, recently, researchers have focused on supervised end-to-end
hybrid approaches, which contain an extractor module and abstractor module.
Among them, the extractor identifies the salient sentences from the input
document, and the abstractor generates a summary from the salient sentences.
This model successfully keeps the consistency between the generated summary and
the reference summary via various strategies (e.g., reinforcement learning).
There are two semantic gaps when training the hybrid model (one is between
document and extracted sentences, and the other is between extracted sentences
and summary). However, they are not explicitly considered in the existing
methods, which usually results in a semantic bias of summary. To mitigate the
above issue, in this paper, a new \textbf{r}einforcing
s\textbf{e}mantic-\textbf{sy}mmetry learning \textbf{m}odel is proposed for
document summarization (\textbf{ReSyM}). ReSyM introduces a
semantic-consistency reward in the extractor to bridge the first gap. A
semantic dual-reward is designed to bridge the second gap in the abstractor.
The whole document summarization process is implemented via reinforcement
learning with a hybrid reward mechanism (combining the above two rewards).
Moreover, a comprehensive sentence representation learning method is presented
to sufficiently capture the information from the original document. A series of
experiments have been conducted on two wildly used benchmark datasets CNN/Daily
Mail and BigPatent. The results have shown the superiority of ReSyM by
comparing it with the state-of-the-art baselines in terms of various evaluation
metrics.
Related papers
- Thesis: Document Summarization with applications to Keyword extraction and Image Retrieval [0.0]
We propose a set of submodular functions for opinion summarization.
Opinion summarization has built in it the tasks of summarization and sentiment detection.
Our functions generate summaries such as there is good correlation between document sentiment and summary sentiment along with good ROUGE score.
arXiv Detail & Related papers (2024-05-20T21:27:18Z) - Source Identification in Abstractive Summarization [0.8883733362171033]
We define input sentences that contain essential information in the generated summary as $textitsource sentences$ and study how abstractive summaries are made by analyzing the source sentences.
We formulate automatic source sentence detection and compare multiple methods to establish a strong baseline for the task.
Experimental results show that the perplexity-based method performs well in highly abstractive settings, while similarity-based methods robustly in relatively extractive settings.
arXiv Detail & Related papers (2024-02-07T09:09:09Z) - On Context Utilization in Summarization with Large Language Models [83.84459732796302]
Large language models (LLMs) excel in abstractive summarization tasks, delivering fluent and pertinent summaries.
Recent advancements have extended their capabilities to handle long-input contexts, exceeding 100k tokens.
We conduct the first comprehensive study on context utilization and position bias in summarization.
arXiv Detail & Related papers (2023-10-16T16:45:12Z) - Text Summarization with Oracle Expectation [88.39032981994535]
Extractive summarization produces summaries by identifying and concatenating the most important sentences in a document.
Most summarization datasets do not come with gold labels indicating whether document sentences are summary-worthy.
We propose a simple yet effective labeling algorithm that creates soft, expectation-based sentence labels.
arXiv Detail & Related papers (2022-09-26T14:10:08Z) - An Efficient Coarse-to-Fine Facet-Aware Unsupervised Summarization
Framework based on Semantic Blocks [27.895044398724664]
We propose an efficient Coarse-to-Fine Facet-Aware Ranking (C2F-FAR) framework for unsupervised long document summarization.
In the coarse-level stage, we propose a new segment algorithm to split the document into facet-aware semantic blocks and then filter insignificant blocks.
In the fine-level stage, we select salient sentences in each block and then extract the final summary from selected sentences.
arXiv Detail & Related papers (2022-08-17T12:18:36Z) - SgSum: Transforming Multi-document Summarization into Sub-graph
Selection [27.40759123902261]
Most existing extractive multi-document summarization (MDS) methods score each sentence individually and extract salient sentences one by one to compose a summary.
We propose a novel MDS framework (SgSum) to formulate the MDS task as a sub-graph selection problem.
Our model can produce significantly more coherent and informative summaries compared with traditional MDS methods.
arXiv Detail & Related papers (2021-10-25T05:12:10Z) - Unsupervised Extractive Summarization by Pre-training Hierarchical
Transformers [107.12125265675483]
Unsupervised extractive document summarization aims to select important sentences from a document without using labeled summaries during training.
Existing methods are mostly graph-based with sentences as nodes and edge weights measured by sentence similarities.
We find that transformer attentions can be used to rank sentences for unsupervised extractive summarization.
arXiv Detail & Related papers (2020-10-16T08:44:09Z) - Extractive Summarization as Text Matching [123.09816729675838]
This paper creates a paradigm shift with regard to the way we build neural extractive summarization systems.
We formulate the extractive summarization task as a semantic text matching problem.
We have driven the state-of-the-art extractive result on CNN/DailyMail to a new level (44.41 in ROUGE-1)
arXiv Detail & Related papers (2020-04-19T08:27:57Z) - At Which Level Should We Extract? An Empirical Analysis on Extractive
Document Summarization [110.54963847339775]
We show that unnecessity and redundancy issues exist when extracting full sentences.
We propose extracting sub-sentential units based on the constituency parsing tree.
arXiv Detail & Related papers (2020-04-06T13:35:10Z) - Pre-training for Abstractive Document Summarization by Reinstating
Source Text [105.77348528847337]
This paper presents three pre-training objectives which allow us to pre-train a Seq2Seq based abstractive summarization model on unlabeled text.
Experiments on two benchmark summarization datasets show that all three objectives can improve performance upon baselines.
arXiv Detail & Related papers (2020-04-04T05:06:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.