Minuteman: Machine and Human Joining Forces in Meeting Summarization
- URL: http://arxiv.org/abs/2309.05272v1
- Date: Mon, 11 Sep 2023 07:10:47 GMT
- Title: Minuteman: Machine and Human Joining Forces in Meeting Summarization
- Authors: Franti\v{s}ek Kmje\v{c}, Ond\v{r}ej Bojar
- Abstract summary: We propose a novel tool to enable efficient semi-automatic meeting minuting.
The tool provides a live transcript and a live meeting summary to the users, who can edit them in a collaborative manner.
The resulting application eases the cognitive load of the notetakers and allows them to easily catch up if they missed a part of the meeting due to absence or a lack of focus.
- Score: 2.900810893770134
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many meetings require creating a meeting summary to keep everyone up to date.
Creating minutes of sufficient quality is however very cognitively demanding.
Although we currently possess capable models for both audio speech recognition
(ASR) and summarization, their fully automatic use is still problematic. ASR
models frequently commit errors when transcribing named entities while the
summarization models tend to hallucinate and misinterpret the transcript. We
propose a novel tool -- Minuteman -- to enable efficient semi-automatic meeting
minuting. The tool provides a live transcript and a live meeting summary to the
users, who can edit them in a collaborative manner, enabling correction of ASR
errors and imperfect summary points in real time. The resulting application
eases the cognitive load of the notetakers and allows them to easily catch up
if they missed a part of the meeting due to absence or a lack of focus. We
conduct several tests of the application in varied settings, exploring the
worthiness of the concept and the possible user strategies.
Related papers
- Predictive Speech Recognition and End-of-Utterance Detection Towards Spoken Dialog Systems [55.99999020778169]
We study a function that can predict the forthcoming words and estimate the time remaining until the end of an utterance.
We develop a cross-attention-based algorithm that incorporates both acoustic and linguistic information.
Results demonstrate the proposed model's ability to predict upcoming words and estimate future EOU events up to 300ms prior to the actual EOU.
arXiv Detail & Related papers (2024-09-30T06:29:58Z) - What's Wrong? Refining Meeting Summaries with LLM Feedback [6.532478490187084]
We introduce a multi-LLM correction approach for meeting summarization using a two-phase process that mimics the human review process.
We release QMSum Mistake, a dataset of 200 automatically generated meeting summaries annotated by humans on nine error types.
We transform identified mistakes into actionable feedback to improve the quality of a given summary measured by relevance, informativeness, conciseness, and coherence.
arXiv Detail & Related papers (2024-07-16T17:10:16Z) - PRoDeliberation: Parallel Robust Deliberation for End-to-End Spoken Language Understanding [44.77985942208969]
PRoDeliberation is a novel method leveraging a Connectionist Temporal Classification-based decoding strategy as well as a denoising objective to train robust non-autoregressive deliberation models.
We show that PRoDeliberation achieves the latency reduction of parallel decoding (2-10x improvement over autoregressive models) while retaining the ability to correct Automatic Speech Recognition (ASR) mistranscriptions.
arXiv Detail & Related papers (2024-06-12T02:46:17Z) - From Instructions to Constraints: Language Model Alignment with
Automatic Constraint Verification [70.08146540745877]
We investigate common constraints in NLP tasks, categorize them into three classes based on the types of their arguments.
We propose a unified framework, ACT (Aligning to ConsTraints), to automatically produce supervision signals for user alignment with constraints.
arXiv Detail & Related papers (2024-03-10T22:14:54Z) - Summaries, Highlights, and Action items: Design, implementation and evaluation of an LLM-powered meeting recap system [30.35387091657807]
Large language models (LLMs) for dialog summarization have the potential to improve the experience of meetings.
Despite this potential, they face technological limitation due to long transcripts and inability to capture diverse recap needs based on user's context.
We develop a system to operationalize the representations with dialogue summarization as its building blocks.
arXiv Detail & Related papers (2023-07-28T20:25:11Z) - A Sliding-Window Approach to Automatic Creation of Meeting Minutes [66.39584679676817]
Meeting minutes record any subject matters discussed, decisions reached and actions taken at meetings.
We present a sliding window approach to automatic generation of meeting minutes.
It aims to tackle issues associated with the nature of spoken text, including lengthy transcripts and lack of document structure.
arXiv Detail & Related papers (2021-04-26T02:44:14Z) - Knowledge Distillation for Improved Accuracy in Spoken Question
Answering [63.72278693825945]
We devise a training strategy to perform knowledge distillation from spoken documents and written counterparts.
Our work makes a step towards distilling knowledge from the language model as a supervision signal.
Experiments demonstrate that our approach outperforms several state-of-the-art language models on the Spoken-SQuAD dataset.
arXiv Detail & Related papers (2020-10-21T15:18:01Z) - Improving Readability for Automatic Speech Recognition Transcription [50.86019112545596]
We propose a novel NLP task called ASR post-processing for readability (APR)
APR aims to transform the noisy ASR output into a readable text for humans and downstream tasks while maintaining the semantic meaning of the speaker.
We compare fine-tuned models based on several open-sourced and adapted pre-trained models with the traditional pipeline method.
arXiv Detail & Related papers (2020-04-09T09:26:42Z) - A Hierarchical Network for Abstractive Meeting Summarization with
Cross-Domain Pretraining [52.11221075687124]
We propose a novel abstractive summary network that adapts to the meeting scenario.
We design a hierarchical structure to accommodate long meeting transcripts and a role vector to depict the difference among speakers.
Our model outperforms previous approaches in both automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-04-04T21:00:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.