Query-Utterance Attention with Joint modeling for Query-Focused Meeting
Summarization
- URL: http://arxiv.org/abs/2303.04487v3
- Date: Thu, 1 Jun 2023 03:07:31 GMT
- Title: Query-Utterance Attention with Joint modeling for Query-Focused Meeting
Summarization
- Authors: Xingxian Liu, Bin Duan, Bo Xiao, Yajing Xu
- Abstract summary: We propose a query-aware framework with joint modeling token and utterance based on Query-Utterance Attention.
We show that the query relevance of different granularities contributes to generating a summary more related to the query.
- Score: 4.763356598070365
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Query-focused meeting summarization (QFMS) aims to generate summaries from
meeting transcripts in response to a given query. Previous works typically
concatenate the query with meeting transcripts and implicitly model the query
relevance only at the token level with attention mechanism. However, due to the
dilution of key query-relevant information caused by long meeting transcripts,
the original transformer-based model is insufficient to highlight the key parts
related to the query. In this paper, we propose a query-aware framework with
joint modeling token and utterance based on Query-Utterance Attention. It
calculates the utterance-level relevance to the query with a dense retrieval
module. Then both token-level query relevance and utterance-level query
relevance are combined and incorporated into the generation process with
attention mechanism explicitly. We show that the query relevance of different
granularities contributes to generating a summary more related to the query.
Experimental results on the QMSum dataset show that the proposed model achieves
new state-of-the-art performance.
Related papers
- Database-Augmented Query Representation for Information Retrieval [59.57065228857247]
We present a novel retrieval framework called Database-Augmented Query representation (DAQu)
DAQu augments the original query with various (query-related) metadata across multiple tables.
We validate DAQu in diverse retrieval scenarios that can incorporate metadata from the relational database.
arXiv Detail & Related papers (2024-06-23T05:02:21Z) - Selecting Query-bag as Pseudo Relevance Feedback for Information-seeking Conversations [76.70349332096693]
Information-seeking dialogue systems are widely used in e-commerce systems.
We propose a Query-bag based Pseudo Relevance Feedback framework (QB-PRF)
It constructs a query-bag with related queries to serve as pseudo signals to guide information-seeking conversations.
arXiv Detail & Related papers (2024-03-22T08:10:32Z) - Investigating Consistency in Query-Based Meeting Summarization: A
Comparative Study of Different Embedding Methods [0.0]
Text Summarization is one of famous applications in Natural Language Processing (NLP) field.
It aims to automatically generate summary with important information based on a given context.
In this paper, we are inspired by "QMSum: A New Benchmark for Query-based Multi-domain Meeting Summarization" proposed by Microsoft.
We also propose our Locater model designed to extract relevant spans based on given transcript and query, which are then summarized by Summarizer model.
arXiv Detail & Related papers (2024-02-10T08:25:30Z) - Improving Query-Focused Meeting Summarization with Query-Relevant
Knowledge [71.14873115781366]
We propose a knowledge-enhanced two-stage framework called Knowledge-Aware Summarizer (KAS) to tackle the challenges.
In the first stage, we introduce knowledge-aware scores to improve the query-relevant segment extraction.
In the second stage, we incorporate query-relevant knowledge in the summary generation.
arXiv Detail & Related papers (2023-09-05T10:26:02Z) - Learning to Rank Utterances for Query-Focused Meeting Summarization [0.7868449549351486]
We propose a Ranker-Generator framework to rank utterances.
We show that learning to rank utterances helps to select utterances related to the query effectively.
Experimental results on QMSum show that the proposed model outperforms all existing multi-stage models with fewer parameters.
arXiv Detail & Related papers (2023-05-22T06:25:09Z) - CAPSTONE: Curriculum Sampling for Dense Retrieval with Document
Expansion [68.19934563919192]
We propose a curriculum sampling strategy that utilizes pseudo queries during training and progressively enhances the relevance between the generated query and the real query.
Experimental results on both in-domain and out-of-domain datasets demonstrate that our approach outperforms previous dense retrieval models.
arXiv Detail & Related papers (2022-12-18T15:57:46Z) - Few-shot Query-Focused Summarization with Prefix-Merging [10.572282987037353]
We propose a prefix-based pretraining strategy for few-shot learning in query-focused summarization.
With only a small amount of trainable parameters, prefix-merging outperforms fine-tuning on query-focused summarization.
arXiv Detail & Related papers (2022-11-29T12:48:37Z) - Improve Query Focused Abstractive Summarization by Incorporating Answer
Relevance [43.820971952979875]
We propose QFS-BART, a model that incorporates the explicit answer relevance of the source documents given the query via a question answering model.
Our model can take advantage of large pre-trained models which improve the summarization performance significantly.
Empirical results on the Debatepedia dataset show that the proposed model achieves the new state-of-the-art performance.
arXiv Detail & Related papers (2021-05-27T06:58:42Z) - Improving Attention Mechanism with Query-Value Interaction [92.67156911466397]
We propose a query-value interaction function which can learn query-aware attention values.
Our approach can consistently improve the performance of many attention-based models.
arXiv Detail & Related papers (2020-10-08T05:12:52Z) - Query Focused Multi-Document Summarization with Distant Supervision [88.39032981994535]
Existing work relies heavily on retrieval-style methods for estimating the relevance between queries and text segments.
We propose a coarse-to-fine modeling framework which introduces separate modules for estimating whether segments are relevant to the query.
We demonstrate that our framework outperforms strong comparison systems on standard QFS benchmarks.
arXiv Detail & Related papers (2020-04-06T22:35:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.