Unsupervised Summarization with Customized Granularities
- URL: http://arxiv.org/abs/2201.12502v1
- Date: Sat, 29 Jan 2022 05:56:35 GMT
- Title: Unsupervised Summarization with Customized Granularities
- Authors: Ming Zhong, Yang Liu, Suyu Ge, Yuning Mao, Yizhu Jiao, Xingxing Zhang,
Yichong Xu, Chenguang Zhu, Michael Zeng, Jiawei Han
- Abstract summary: We propose the first unsupervised multi-granularity summarization framework, GranuSum.
By inputting different numbers of events, GranuSum is capable of producing multi-granular summaries in an unsupervised manner.
- Score: 76.26899748972423
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Text summarization is a personalized and customized task, i.e., for one
document, users often have different preferences for the summary. As a key
aspect of customization in summarization, granularity is used to measure the
semantic coverage between summary and source document. Coarse-grained summaries
can only contain the most central event in the original text, while
fine-grained summaries cover more sub-events and corresponding details.
However, previous studies mostly develop systems in the single-granularity
scenario. And models that can generate summaries with customizable semantic
coverage still remain an under-explored topic. In this paper, we propose the
first unsupervised multi-granularity summarization framework, GranuSum. We take
events as the basic semantic units of the source documents and propose to rank
these events by their salience. We also develop a model to summarize input
documents with given events as anchors and hints. By inputting different
numbers of events, GranuSum is capable of producing multi-granular summaries in
an unsupervised manner. Meanwhile, to evaluate multi-granularity summarization
models, we annotate a new benchmark GranuDUC, in which we write multiple
summaries of different granularities for each document cluster. Experimental
results confirm the substantial superiority of GranuSum on multi-granularity
summarization over several baseline systems. Furthermore, by experimenting on
conventional unsupervised abstractive summarization tasks, we find that
GranuSum, by exploiting the event information, can also achieve new
state-of-the-art results under this scenario, outperforming strong baselines.
Related papers
- GLIMMER: Incorporating Graph and Lexical Features in Unsupervised Multi-Document Summarization [13.61818620609812]
We propose a lightweight yet effective unsupervised approach called GLIMMER: a Graph and LexIcal features based unsupervised Multi-docuMEnt summaRization approach.
It first constructs a sentence graph from the source documents, then automatically identifies semantic clusters by mining low-level features from raw texts.
Experiments conducted on Multi-News, Multi-XScience and DUC-2004 demonstrate that our approach outperforms existing unsupervised approaches.
arXiv Detail & Related papers (2024-08-19T16:01:48Z) - Write Summary Step-by-Step: A Pilot Study of Stepwise Summarization [48.57273563299046]
We propose the task of Stepwise Summarization, which aims to generate a new appended summary each time a new document is proposed.
The appended summary should not only summarize the newly added content but also be coherent with the previous summary.
We show that SSG achieves state-of-the-art performance in terms of both automatic metrics and human evaluations.
arXiv Detail & Related papers (2024-06-08T05:37:26Z) - JADS: A Framework for Self-supervised Joint Aspect Discovery and Summarization [3.992091862806936]
Our solution integrates topic discovery and summarization into a single step.
Given text data, our Joint Aspect Discovery and Summarization algorithm (JADS) discovers aspects from the input.
Our proposed method achieves higher semantic alignment with ground truth and is factual.
arXiv Detail & Related papers (2024-05-28T23:01:57Z) - Event-Keyed Summarization [46.521305453350635]
Event-keyed summarization (EKS) is a novel task that marries traditional summarization and document-level event extraction.
We introduce a dataset for this task, MUCSUM, consisting of summaries of all events in the classic MUC-4 dataset.
We show that ablations that reduce EKS to traditional summarization or structure-to-text yield inferior summaries of target events.
arXiv Detail & Related papers (2024-02-10T15:32:53Z) - UniSumm and SummZoo: Unified Model and Diverse Benchmark for Few-Shot
Summarization [54.59104881168188]
textscUniSumm is a unified few-shot summarization model pre-trained with multiple summarization tasks.
textscSummZoo is a new benchmark to better evaluate few-shot summarizers.
arXiv Detail & Related papers (2022-11-17T18:54:47Z) - Zero-Shot On-the-Fly Event Schema Induction [61.91468909200566]
We present a new approach in which large language models are utilized to generate source documents that allow predicting, given a high-level event definition, the specific events, arguments, and relations between them.
Using our model, complete schemas on any topic can be generated on-the-fly without any manual data collection, i.e., in a zero-shot manner.
arXiv Detail & Related papers (2022-10-12T14:37:00Z) - Text Summarization with Latent Queries [60.468323530248945]
We introduce LaQSum, the first unified text summarization system that learns Latent Queries from documents for abstractive summarization with any existing query forms.
Under a deep generative framework, our system jointly optimize a latent query model and a conditional language model, allowing users to plug-and-play queries of any type at test time.
Our system robustly outperforms strong comparison systems across summarization benchmarks with different query types, document settings, and target domains.
arXiv Detail & Related papers (2021-05-31T21:14:58Z) - An Enhanced MeanSum Method For Generating Hotel Multi-Review
Summarizations [0.06091702876917279]
This work uses Multi-Aspect Masker(MAM) as content selector to address the issue with multi-aspect.
We also propose a regularizer to control the length of the generated summaries.
Our improved model achieves higher ROUGE, Sentiment Accuracy than the original Meansum method.
arXiv Detail & Related papers (2020-12-07T13:16:01Z) - SupMMD: A Sentence Importance Model for Extractive Summarization using
Maximum Mean Discrepancy [92.5683788430012]
SupMMD is a novel technique for generic and update summarization based on the maximum discrepancy from kernel two-sample testing.
We show the efficacy of SupMMD in both generic and update summarization tasks by meeting or exceeding the current state-of-the-art on the DUC-2004 and TAC-2009 datasets.
arXiv Detail & Related papers (2020-10-06T09:26:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.