Extractive Summarization of Call Transcripts
- URL: http://arxiv.org/abs/2103.10599v1
- Date: Fri, 19 Mar 2021 02:40:59 GMT
- Title: Extractive Summarization of Call Transcripts
- Authors: Pratik K. Biswas and Aleksandr Iakubovich
- Abstract summary: This paper presents an indigenously developed method that combines topic modeling and sentence selection with punctuation restoration in ill-punctuated or un-punctuated call transcripts.
Extensive testing, evaluation and comparisons have demonstrated the efficacy of this summarizer for call transcript summarization.
- Score: 77.96603959765577
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Text summarization is the process of extracting the most important
information from the text and presenting it concisely in fewer sentences. Call
transcript is a text that involves textual description of a phone conversation
between a customer (caller) and agent(s) (customer representatives). This paper
presents an indigenously developed method that combines topic modeling and
sentence selection with punctuation restoration in condensing ill-punctuated or
un-punctuated call transcripts to produce summaries that are more readable.
Extensive testing, evaluation and comparisons have demonstrated the efficacy of
this summarizer for call transcript summarization.
Related papers
- Speech vs. Transcript: Does It Matter for Human Annotators in Speech Summarization? [35.71047777304832]
We examine whether summaries based on annotators listening to the recordings differ from those based on annotators reading transcripts.
We find that summaries are indeed different based on the source modality, and that speech-based summaries are more factually consistent and information-selective than transcript-based summaries.
arXiv Detail & Related papers (2024-08-12T13:25:53Z) - `Keep it Together': Enforcing Cohesion in Extractive Summaries by
Simulating Human Memory [22.659031563705245]
In this paper, we aim to enforce cohesion whilst controlling for informativeness and redundancy in summaries.
Our sentence selector simulates human memory to keep track of topics.
It is possible to extract highly cohesive summaries that nevertheless read as informative to humans.
arXiv Detail & Related papers (2024-02-16T12:43:26Z) - A Hybrid Strategy for Chat Transcript Summarization [3.0567294793102784]
Sumomarization is the process of condensing a piece of text to fewer sentences, while still preserving its content.
This paper presents an indigenously developed hybrid method that combines extractive and abstractive summarization techniques.
Tests, evaluations, comparisons, and validation have demonstrated the efficacy of this approach for large-scale deployment of chat transcript summarization.
arXiv Detail & Related papers (2024-02-02T15:44:28Z) - Factually Consistent Summarization via Reinforcement Learning with
Textual Entailment Feedback [57.816210168909286]
We leverage recent progress on textual entailment models to address this problem for abstractive summarization systems.
We use reinforcement learning with reference-free, textual entailment rewards to optimize for factual consistency.
Our results, according to both automatic metrics and human evaluation, show that our method considerably improves the faithfulness, salience, and conciseness of the generated summaries.
arXiv Detail & Related papers (2023-05-31T21:04:04Z) - Toward Unifying Text Segmentation and Long Document Summarization [31.084738269628748]
We study the role that section segmentation plays in extractive summarization of written and spoken documents.
Our approach learns robust sentence representations by performing summarization and segmentation simultaneously.
Our findings suggest that the model can not only achieve state-of-the-art performance on publicly available benchmarks, but demonstrate better cross-genre transferability.
arXiv Detail & Related papers (2022-10-28T22:07:10Z) - Comparing Methods for Extractive Summarization of Call Centre Dialogue [77.34726150561087]
We experimentally compare several such methods by using them to produce summaries of calls, and evaluating these summaries objectively.
We found that TopicSum and Lead-N outperform the other summarisation methods, whilst BERTSum received comparatively lower scores in both subjective and objective evaluations.
arXiv Detail & Related papers (2022-09-06T13:16:02Z) - A Survey on Neural Abstractive Summarization Methods and Factual
Consistency of Summarization [18.763290930749235]
summarization is the process of shortening a set of textual data computationally, to create a subset (a summary)
Existing summarization methods can be roughly divided into two types: extractive and abstractive.
An extractive summarizer explicitly selects text snippets from the source document, while an abstractive summarizer generates novel text snippets to convey the most salient concepts prevalent in the source.
arXiv Detail & Related papers (2022-04-20T14:56:36Z) - A Sliding-Window Approach to Automatic Creation of Meeting Minutes [66.39584679676817]
Meeting minutes record any subject matters discussed, decisions reached and actions taken at meetings.
We present a sliding window approach to automatic generation of meeting minutes.
It aims to tackle issues associated with the nature of spoken text, including lengthy transcripts and lack of document structure.
arXiv Detail & Related papers (2021-04-26T02:44:14Z) - Relation Clustering in Narrative Knowledge Graphs [71.98234178455398]
relational sentences in the original text are embedded (with SBERT) and clustered in order to merge together semantically similar relations.
Preliminary tests show that such clustering might successfully detect similar relations, and provide a valuable preprocessing for semi-supervised approaches.
arXiv Detail & Related papers (2020-11-27T10:43:04Z) - Screenplay Summarization Using Latent Narrative Structure [78.45316339164133]
We propose to explicitly incorporate the underlying structure of narratives into general unsupervised and supervised extractive summarization models.
We formalize narrative structure in terms of key narrative events (turning points) and treat it as latent in order to summarize screenplays.
Experimental results on the CSI corpus of TV screenplays, which we augment with scene-level summarization labels, show that latent turning points correlate with important aspects of a CSI episode.
arXiv Detail & Related papers (2020-04-27T11:54:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.