Topic-Guided Abstractive Multi-Document Summarization
- URL: http://arxiv.org/abs/2110.11207v1
- Date: Thu, 21 Oct 2021 15:32:30 GMT
- Title: Topic-Guided Abstractive Multi-Document Summarization
- Authors: Peng Cui, Le Hu
- Abstract summary: A critical point of multi-document summarization (MDS) is to learn the relations among various documents.
We propose a novel abstractive MDS model, in which we represent multiple documents as a heterogeneous graph.
We employ a neural topic model to jointly discover latent topics that can act as cross-document semantic units.
- Score: 21.856615677793243
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A critical point of multi-document summarization (MDS) is to learn the
relations among various documents. In this paper, we propose a novel
abstractive MDS model, in which we represent multiple documents as a
heterogeneous graph, taking semantic nodes of different granularities into
account, and then apply a graph-to-sequence framework to generate summaries.
Moreover, we employ a neural topic model to jointly discover latent topics that
can act as cross-document semantic units to bridge different documents and
provide global information to guide the summary generation. Since topic
extraction can be viewed as a special type of summarization that "summarizes"
texts into a more abstract format, i.e., a topic distribution, we adopt a
multi-task learning strategy to jointly train the topic and summarization
module, allowing the promotion of each other. Experimental results on the
Multi-News dataset demonstrate that our model outperforms previous
state-of-the-art MDS models on both Rouge metrics and human evaluation,
meanwhile learns high-quality topics.
Related papers
- Peek Across: Improving Multi-Document Modeling via Cross-Document
Question-Answering [49.85790367128085]
We pre-training a generic multi-document model from a novel cross-document question answering pre-training objective.
This novel multi-document QA formulation directs the model to better recover cross-text informational relations.
Unlike prior multi-document models that focus on either classification or summarization tasks, our pre-training objective formulation enables the model to perform tasks that involve both short text generation and long text generation.
arXiv Detail & Related papers (2023-05-24T17:48:40Z) - Information Screening whilst Exploiting! Multimodal Relation Extraction
with Feature Denoising and Multimodal Topic Modeling [96.75821232222201]
Existing research on multimodal relation extraction (MRE) faces two co-existing challenges, internal-information over-utilization and external-information under-exploitation.
We propose a novel framework that simultaneously implements the idea of internal-information screening and external-information exploiting.
arXiv Detail & Related papers (2023-05-19T14:56:57Z) - Moving beyond word lists: towards abstractive topic labels for
human-like topics of scientific documents [0.0]
We present an approach to generating human-like topic labels using abstractive multi-document summarization (MDS)
We model topics in citation sentences in order to understand what further research needs to be done to fully operationalize MDS for topic labeling.
arXiv Detail & Related papers (2022-10-28T17:47:12Z) - Unsupervised Summarization with Customized Granularities [76.26899748972423]
We propose the first unsupervised multi-granularity summarization framework, GranuSum.
By inputting different numbers of events, GranuSum is capable of producing multi-granular summaries in an unsupervised manner.
arXiv Detail & Related papers (2022-01-29T05:56:35Z) - Modeling Endorsement for Multi-Document Abstractive Summarization [10.166639983949887]
A crucial difference between single- and multi-document summarization is how salient content manifests itself in the document(s)
In this paper, we model the cross-document endorsement effect and its utilization in multiple document summarization.
Our method generates a synopsis from each document, which serves as an endorser to identify salient content from other documents.
arXiv Detail & Related papers (2021-10-15T03:55:42Z) - Multi-document Summarization via Deep Learning Techniques: A Survey [29.431160110691607]
We propose a novel taxonomy to summarize the design strategies of neural networks.
We highlight the differences between various objective functions that are rarely discussed in the existing literature.
arXiv Detail & Related papers (2020-11-10T00:35:46Z) - Topic-Guided Abstractive Text Summarization: a Joint Learning Approach [19.623946402970933]
We introduce a new approach for abstractive text summarization, Topic-Guided Abstractive Summarization.
The idea is to incorporate neural topic modeling with a Transformer-based sequence-to-sequence (seq2seq) model in a joint learning framework.
arXiv Detail & Related papers (2020-10-20T14:45:25Z) - SupMMD: A Sentence Importance Model for Extractive Summarization using
Maximum Mean Discrepancy [92.5683788430012]
SupMMD is a novel technique for generic and update summarization based on the maximum discrepancy from kernel two-sample testing.
We show the efficacy of SupMMD in both generic and update summarization tasks by meeting or exceeding the current state-of-the-art on the DUC-2004 and TAC-2009 datasets.
arXiv Detail & Related papers (2020-10-06T09:26:55Z) - Topic Adaptation and Prototype Encoding for Few-Shot Visual Storytelling [81.33107307509718]
We propose a topic adaptive storyteller to model the ability of inter-topic generalization.
We also propose a prototype encoding structure to model the ability of intra-topic derivation.
Experimental results show that topic adaptation and prototype encoding structure mutually bring benefit to the few-shot model.
arXiv Detail & Related papers (2020-08-11T03:55:11Z) - Leveraging Graph to Improve Abstractive Multi-Document Summarization [50.62418656177642]
We develop a neural abstractive multi-document summarization (MDS) model which can leverage well-known graph representations of documents.
Our model utilizes graphs to encode documents in order to capture cross-document relations, which is crucial to summarizing long documents.
Our model can also take advantage of graphs to guide the summary generation process, which is beneficial for generating coherent and concise summaries.
arXiv Detail & Related papers (2020-05-20T13:39:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.