A Hierarchical Network for Abstractive Meeting Summarization with
Cross-Domain Pretraining
- URL: http://arxiv.org/abs/2004.02016v4
- Date: Sun, 20 Sep 2020 05:47:23 GMT
- Title: A Hierarchical Network for Abstractive Meeting Summarization with
Cross-Domain Pretraining
- Authors: Chenguang Zhu, Ruochen Xu, Michael Zeng, Xuedong Huang
- Abstract summary: We propose a novel abstractive summary network that adapts to the meeting scenario.
We design a hierarchical structure to accommodate long meeting transcripts and a role vector to depict the difference among speakers.
Our model outperforms previous approaches in both automatic metrics and human evaluation.
- Score: 52.11221075687124
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the abundance of automatic meeting transcripts, meeting summarization is
of great interest to both participants and other parties. Traditional methods
of summarizing meetings depend on complex multi-step pipelines that make joint
optimization intractable. Meanwhile, there are a handful of deep neural models
for text summarization and dialogue systems. However, the semantic structure
and styles of meeting transcripts are quite different from articles and
conversations. In this paper, we propose a novel abstractive summary network
that adapts to the meeting scenario. We design a hierarchical structure to
accommodate long meeting transcripts and a role vector to depict the difference
among speakers. Furthermore, due to the inadequacy of meeting summary data, we
pretrain the model on large-scale news summary data. Empirical results show
that our model outperforms previous approaches in both automatic metrics and
human evaluation. For example, on ICSI dataset, the ROUGE-1 score increases
from 34.66% to 46.28%.
Related papers
- Leveraging Discourse Structure for Extractive Meeting Summarization [26.76383031532945]
We introduce an extractive summarization system for meetings that leverages discourse structure to better identify salient information.
We train a GNN-based node classification model to select the most important utterances, which are then combined to create an extractive summary.
Experimental results on AMI and ICSI demonstrate that our approach surpasses existing text-based and graph-based extractive summarization systems.
arXiv Detail & Related papers (2024-05-17T19:06:20Z) - Investigating Consistency in Query-Based Meeting Summarization: A
Comparative Study of Different Embedding Methods [0.0]
Text Summarization is one of famous applications in Natural Language Processing (NLP) field.
It aims to automatically generate summary with important information based on a given context.
In this paper, we are inspired by "QMSum: A New Benchmark for Query-based Multi-domain Meeting Summarization" proposed by Microsoft.
We also propose our Locater model designed to extract relevant spans based on given transcript and query, which are then summarized by Summarizer model.
arXiv Detail & Related papers (2024-02-10T08:25:30Z) - Action-Item-Driven Summarization of Long Meeting Transcripts [8.430481660019451]
This paper introduces a novel and effective approach to automate the generation of meeting summaries.
Our novel algorithms can generate abstractive meeting summaries that are driven by the action items contained in the meeting transcript.
Our pipeline achieved a BERTScore of 64.98 across the AMI corpus, which is an approximately 4.98% increase from the current state-of-the-art result.
arXiv Detail & Related papers (2023-12-29T12:33:21Z) - DIONYSUS: A Pre-trained Model for Low-Resource Dialogue Summarization [127.714919036388]
DIONYSUS is a pre-trained encoder-decoder model for summarizing dialogues in any new domain.
Our experiments show that DIONYSUS outperforms existing methods on six datasets.
arXiv Detail & Related papers (2022-12-20T06:21:21Z) - TANet: Thread-Aware Pretraining for Abstractive Conversational
Summarization [27.185068253347257]
We build a large-scale (11M) pretraining dataset called RCS based on the multi-person discussions in the Reddit community.
We then present TANet, a thread-aware Transformer-based network.
Unlike the existing pre-trained models that treat a conversation as a sequence of sentences, we argue that the inherent contextual dependency plays an essential role in understanding the entire conversation.
arXiv Detail & Related papers (2022-04-09T16:08:46Z) - Long Document Summarization with Top-down and Bottom-up Inference [113.29319668246407]
We propose a principled inference framework to improve summarization models on two aspects.
Our framework assumes a hierarchical latent structure of a document where the top-level captures the long range dependency.
We demonstrate the effectiveness of the proposed framework on a diverse set of summarization datasets.
arXiv Detail & Related papers (2022-03-15T01:24:51Z) - ConvoSumm: Conversation Summarization Benchmark and Improved Abstractive
Summarization with Argument Mining [61.82562838486632]
We crowdsource four new datasets on diverse online conversation forms of news comments, discussion forums, community question answering forums, and email threads.
We benchmark state-of-the-art models on our datasets and analyze characteristics associated with the data.
arXiv Detail & Related papers (2021-06-01T22:17:13Z) - End-to-End Diarization for Variable Number of Speakers with Local-Global
Networks and Discriminative Speaker Embeddings [66.50782702086575]
We present an end-to-end deep network model that performs meeting diarization from single-channel audio recordings.
The proposed system is designed to handle meetings with unknown numbers of speakers, using variable-number permutation-invariant cross-entropy based loss functions.
arXiv Detail & Related papers (2021-05-05T14:55:29Z) - Dialogue Discourse-Aware Graph Convolutional Networks for Abstractive
Meeting Summarization [24.646506847760822]
We develop a dialogue discourse-Aware Graph Convolutional Networks (DDA-GCN) for meeting summarization.
We first transform the entire meeting text with dialogue discourse relations into a discourse graph and then use DDA-GCN to encode the semantic representation of the graph.
Finally, we employ a Recurrent Neural Network to generate the summary.
arXiv Detail & Related papers (2020-12-07T07:51:38Z) - Multi-View Sequence-to-Sequence Models with Conversational Structure for
Abstractive Dialogue Summarization [72.54873655114844]
Text summarization is one of the most challenging and interesting problems in NLP.
This work proposes a multi-view sequence-to-sequence model by first extracting conversational structures of unstructured daily chats from different views to represent conversations.
Experiments on a large-scale dialogue summarization corpus demonstrated that our methods significantly outperformed previous state-of-the-art models via both automatic evaluations and human judgment.
arXiv Detail & Related papers (2020-10-04T20:12:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.