QMSum: A New Benchmark for Query-based Multi-domain Meeting
Summarization
- URL: http://arxiv.org/abs/2104.05938v1
- Date: Tue, 13 Apr 2021 05:00:35 GMT
- Title: QMSum: A New Benchmark for Query-based Multi-domain Meeting
Summarization
- Authors: Ming Zhong, Da Yin, Tao Yu, Ahmad Zaidi, Mutethia Mutuma, Rahul Jha,
Ahmed Hassan Awadallah, Asli Celikyilmaz, Yang Liu, Xipeng Qiu, Dragomir
Radev
- Abstract summary: QMSum consists of 1,808 query-summary pairs over 232 meetings in multiple domains.
We investigate a locate-then-summarize method and evaluate a set of strong summarization baselines on the task.
- Score: 45.83402681068943
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Meetings are a key component of human collaboration. As increasing numbers of
meetings are recorded and transcribed, meeting summaries have become essential
to remind those who may or may not have attended the meetings about the key
decisions made and the tasks to be completed. However, it is hard to create a
single short summary that covers all the content of a long meeting involving
multiple people and topics. In order to satisfy the needs of different types of
users, we define a new query-based multi-domain meeting summarization task,
where models have to select and summarize relevant spans of meetings in
response to a query, and we introduce QMSum, a new benchmark for this task.
QMSum consists of 1,808 query-summary pairs over 232 meetings in multiple
domains. Besides, we investigate a locate-then-summarize method and evaluate a
set of strong summarization baselines on the task. Experimental results and
manual analysis reveal that QMSum presents significant challenges in long
meeting summarization for future research. Dataset is available at
\url{https://github.com/Yale-LILY/QMSum}.
Related papers
- QFMTS: Generating Query-Focused Summaries over Multi-Table Inputs [63.98556480088152]
Table summarization is a crucial task aimed at condensing information into concise and comprehensible textual summaries.
We propose a novel method to address these limitations by introducing query-focused multi-table summarization.
Our approach, which comprises a table serialization module, a summarization controller, and a large language model, generates query-dependent table summaries tailored to users' information needs.
arXiv Detail & Related papers (2024-05-08T15:05:55Z) - Investigating Consistency in Query-Based Meeting Summarization: A
Comparative Study of Different Embedding Methods [0.0]
Text Summarization is one of famous applications in Natural Language Processing (NLP) field.
It aims to automatically generate summary with important information based on a given context.
In this paper, we are inspired by "QMSum: A New Benchmark for Query-based Multi-domain Meeting Summarization" proposed by Microsoft.
We also propose our Locater model designed to extract relevant spans based on given transcript and query, which are then summarized by Summarizer model.
arXiv Detail & Related papers (2024-02-10T08:25:30Z) - Improving Query-Focused Meeting Summarization with Query-Relevant
Knowledge [71.14873115781366]
We propose a knowledge-enhanced two-stage framework called Knowledge-Aware Summarizer (KAS) to tackle the challenges.
In the first stage, we introduce knowledge-aware scores to improve the query-relevant segment extraction.
In the second stage, we incorporate query-relevant knowledge in the summary generation.
arXiv Detail & Related papers (2023-09-05T10:26:02Z) - VCSUM: A Versatile Chinese Meeting Summarization Dataset [25.695308276427166]
We introduce a versatile Chinese meeting summarization dataset, dubbed VCSum, consisting of 239 real-life meetings.
We provide the annotations of topic segmentation, headlines, segmentation summaries, overall meeting summaries, and salient sentences for each meeting transcript.
Our analysis confirms the effectiveness and robustness of VCSum.
arXiv Detail & Related papers (2023-05-09T09:07:15Z) - Query-Utterance Attention with Joint modeling for Query-Focused Meeting
Summarization [4.763356598070365]
We propose a query-aware framework with joint modeling token and utterance based on Query-Utterance Attention.
We show that the query relevance of different granularities contributes to generating a summary more related to the query.
arXiv Detail & Related papers (2023-03-08T10:21:45Z) - UniSumm and SummZoo: Unified Model and Diverse Benchmark for Few-Shot
Summarization [54.59104881168188]
textscUniSumm is a unified few-shot summarization model pre-trained with multiple summarization tasks.
textscSummZoo is a new benchmark to better evaluate few-shot summarizers.
arXiv Detail & Related papers (2022-11-17T18:54:47Z) - Text Summarization with Latent Queries [60.468323530248945]
We introduce LaQSum, the first unified text summarization system that learns Latent Queries from documents for abstractive summarization with any existing query forms.
Under a deep generative framework, our system jointly optimize a latent query model and a conditional language model, allowing users to plug-and-play queries of any type at test time.
Our system robustly outperforms strong comparison systems across summarization benchmarks with different query types, document settings, and target domains.
arXiv Detail & Related papers (2021-05-31T21:14:58Z) - Abstractive Query Focused Summarization with Query-Free Resources [60.468323530248945]
In this work, we consider the problem of leveraging only generic summarization resources to build an abstractive QFS system.
We propose Marge, a Masked ROUGE Regression framework composed of a novel unified representation for summaries and queries.
Despite learning from minimal supervision, our system achieves state-of-the-art results in the distantly supervised setting.
arXiv Detail & Related papers (2020-12-29T14:39:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.