Centrality Meets Centroid: A Graph-based Approach for Unsupervised
Document Summarization
- URL: http://arxiv.org/abs/2103.15327v1
- Date: Mon, 29 Mar 2021 04:35:33 GMT
- Title: Centrality Meets Centroid: A Graph-based Approach for Unsupervised
Document Summarization
- Authors: Haopeng Zhang and Jiawei Zhang
- Abstract summary: We propose a graph-based unsupervised approach for extractive document summarization.
Our approach works at a summary-level by utilizing graph centrality and centroid.
- Score: 13.12794447731674
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised document summarization has re-acquired lots of attention in
recent years thanks to its simplicity and data independence. In this paper, we
propose a graph-based unsupervised approach for extractive document
summarization. Instead of ranking sentences by salience and extracting
sentences one by one, our approach works at a summary-level by utilizing graph
centrality and centroid. We first extract summary candidates as subgraphs based
on centrality from the sentence graph and then select from the summary
candidates by matching to the centroid. We perform extensive experiments on two
bench-marked summarization datasets, and the results demonstrate the
effectiveness of our model compared to state-of-the-art baselines.
Related papers
- Incremental Extractive Opinion Summarization Using Cover Trees [81.59625423421355]
In online marketplaces user reviews accumulate over time, and opinion summaries need to be updated periodically.
In this work, we study the task of extractive opinion summarization in an incremental setting.
We present an efficient algorithm for accurately computing the CentroidRank summaries in an incremental setting.
arXiv Detail & Related papers (2024-01-16T02:00:17Z) - Scientific Paper Extractive Summarization Enhanced by Citation Graphs [50.19266650000948]
We focus on leveraging citation graphs to improve scientific paper extractive summarization under different settings.
Preliminary results demonstrate that citation graph is helpful even in a simple unsupervised framework.
Motivated by this, we propose a Graph-based Supervised Summarization model (GSS) to achieve more accurate results on the task when large-scale labeled data are available.
arXiv Detail & Related papers (2022-12-08T11:53:12Z) - Unsupervised Extractive Summarization with Heterogeneous Graph
Embeddings for Chinese Document [5.9630342951482085]
We propose an unsupervised extractive summarizaiton method with heterogeneous graph embeddings (HGEs) for Chinese document.
Experimental results demonstrate that our method consistently outperforms the strong baseline in three summarization datasets.
arXiv Detail & Related papers (2022-11-09T06:07:31Z) - An analysis of document graph construction methods for AMR summarization [2.055054374525828]
We present a novel dataset consisting of human-annotated alignments between the nodes of paired documents and summaries.
We apply these two forms of evaluation to prior work as well as a new method for node merging and show that our new method has significantly better performance than prior work.
arXiv Detail & Related papers (2021-11-27T22:12:50Z) - SgSum: Transforming Multi-document Summarization into Sub-graph
Selection [27.40759123902261]
Most existing extractive multi-document summarization (MDS) methods score each sentence individually and extract salient sentences one by one to compose a summary.
We propose a novel MDS framework (SgSum) to formulate the MDS task as a sub-graph selection problem.
Our model can produce significantly more coherent and informative summaries compared with traditional MDS methods.
arXiv Detail & Related papers (2021-10-25T05:12:10Z) - A Training-free and Reference-free Summarization Evaluation Metric via
Centrality-weighted Relevance and Self-referenced Redundancy [60.419107377879925]
We propose a training-free and reference-free summarization evaluation metric.
Our metric consists of a centrality-weighted relevance score and a self-referenced redundancy score.
Our methods can significantly outperform existing methods on both multi-document and single-document summarization evaluation.
arXiv Detail & Related papers (2021-06-26T05:11:27Z) - Unsupervised Extractive Summarization by Pre-training Hierarchical
Transformers [107.12125265675483]
Unsupervised extractive document summarization aims to select important sentences from a document without using labeled summaries during training.
Existing methods are mostly graph-based with sentences as nodes and edge weights measured by sentence similarities.
We find that transformer attentions can be used to rank sentences for unsupervised extractive summarization.
arXiv Detail & Related papers (2020-10-16T08:44:09Z) - Summary-Source Proposition-level Alignment: Task, Datasets and
Supervised Baseline [94.0601799665342]
Aligning sentences in a reference summary with their counterparts in source documents was shown as a useful auxiliary summarization task.
We propose establishing summary-source alignment as an explicit task, while introducing two major novelties.
We create a novel training dataset for proposition-level alignment, derived automatically from available summarization evaluation data.
We present a supervised proposition alignment baseline model, showing improved alignment-quality over the unsupervised approach.
arXiv Detail & Related papers (2020-09-01T17:27:12Z) - SummPip: Unsupervised Multi-Document Summarization with Sentence Graph
Compression [61.97200991151141]
SummPip is an unsupervised method for multi-document summarization.
We convert the original documents to a sentence graph, taking both linguistic and deep representation into account.
We then apply spectral clustering to obtain multiple clusters of sentences, and finally compress each cluster to generate the final summary.
arXiv Detail & Related papers (2020-07-17T13:01:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.