Dialogue Summarization with Supporting Utterance Flow Modeling and Fact
Regularization
- URL: http://arxiv.org/abs/2108.01268v1
- Date: Tue, 3 Aug 2021 03:09:25 GMT
- Title: Dialogue Summarization with Supporting Utterance Flow Modeling and Fact
Regularization
- Authors: Wang Chen, Piji Li, Hou Pong Chan, Irwin King
- Abstract summary: We propose an end-to-end neural model for dialogue summarization with two novel modules.
The supporting utterance flow modeling helps to generate a coherent summary by smoothly shifting the focus from the former utterances to the later ones.
The fact regularization encourages the generated summary to be factually consistent with the ground-truth summary during model training.
- Score: 58.965859508695225
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Dialogue summarization aims to generate a summary that indicates the key
points of a given dialogue. In this work, we propose an end-to-end neural model
for dialogue summarization with two novel modules, namely, the \emph{supporting
utterance flow modeling module} and the \emph{fact regularization module}. The
supporting utterance flow modeling helps to generate a coherent summary by
smoothly shifting the focus from the former utterances to the later ones. The
fact regularization encourages the generated summary to be factually consistent
with the ground-truth summary during model training, which helps to improve the
factual correctness of the generated summary in inference time. Furthermore, we
also introduce a new benchmark dataset for dialogue summarization. Extensive
experiments on both existing and newly-introduced datasets demonstrate the
effectiveness of our model.
Related papers
- Factual Dialogue Summarization via Learning from Large Language Models [35.63037083806503]
Large language model (LLM)-based automatic text summarization models generate more factually consistent summaries.
We employ zero-shot learning to extract symbolic knowledge from LLMs, generating factually consistent (positive) and inconsistent (negative) summaries.
Our approach achieves better factual consistency while maintaining coherence, fluency, and relevance, as confirmed by various automatic evaluation metrics.
arXiv Detail & Related papers (2024-06-20T20:03:37Z) - Information-Theoretic Distillation for Reference-less Summarization [67.51150817011617]
We present a novel framework to distill a powerful summarizer based on the information-theoretic objective for summarization.
We start off from Pythia-2.8B as the teacher model, which is not yet capable of summarization.
We arrive at a compact but powerful summarizer with only 568M parameters that performs competitively against ChatGPT.
arXiv Detail & Related papers (2024-03-20T17:42:08Z) - Baichuan2-Sum: Instruction Finetune Baichuan2-7B Model for Dialogue Summarization [12.45299260235282]
We propose an instruction fine-tuning model: Baichuan2-Sum, for role-oriented diaglouge summarization.
By setting different instructions for different roles, the model can learn from the dialogue interactions and output the expected summaries.
Experiments demonstrate that the proposed model achieves the new state-of-the-art results on two public dialogue summarization datasets.
arXiv Detail & Related papers (2024-01-27T20:20:39Z) - Instructive Dialogue Summarization with Query Aggregations [41.89962538701501]
We introduce instruction-finetuned language models to expand the capability set of dialogue summarization models.
We propose a three-step approach to synthesize high-quality query-based summarization triples.
By training a unified model called InstructDS on three summarization datasets with multi-purpose instructive triples, we expand the capability of dialogue summarization models.
arXiv Detail & Related papers (2023-10-17T04:03:00Z) - DIONYSUS: A Pre-trained Model for Low-Resource Dialogue Summarization [127.714919036388]
DIONYSUS is a pre-trained encoder-decoder model for summarizing dialogues in any new domain.
Our experiments show that DIONYSUS outperforms existing methods on six datasets.
arXiv Detail & Related papers (2022-12-20T06:21:21Z) - He Said, She Said: Style Transfer for Shifting the Perspective of
Dialogues [75.58367095888914]
We define a new style transfer task: perspective shift, which reframes a dialogue from informal first person to a formal third person rephrasing of the text.
As a sample application, we demonstrate that applying perspective shifting to a dialogue summarization dataset (SAMSum) substantially improves the zero-shot performance of extractive news summarization models.
arXiv Detail & Related papers (2022-10-27T14:16:07Z) - Leveraging Non-dialogue Summaries for Dialogue Summarization [1.0742675209112622]
We apply transformations to document summarization data pairs to create training data that better befit dialogue summarization.
We conduct extensive experiments across both English and Korean to verify our approach.
arXiv Detail & Related papers (2022-10-17T23:34:31Z) - SNaC: Coherence Error Detection for Narrative Summarization [73.48220043216087]
We introduce SNaC, a narrative coherence evaluation framework rooted in fine-grained annotations for long summaries.
We develop a taxonomy of coherence errors in generated narrative summaries and collect span-level annotations for 6.6k sentences across 150 book and movie screenplay summaries.
Our work provides the first characterization of coherence errors generated by state-of-the-art summarization models and a protocol for eliciting coherence judgments from crowd annotators.
arXiv Detail & Related papers (2022-05-19T16:01:47Z) - Multi-Fact Correction in Abstractive Text Summarization [98.27031108197944]
Span-Fact is a suite of two factual correction models that leverages knowledge learned from question answering models to make corrections in system-generated summaries via span selection.
Our models employ single or multi-masking strategies to either iteratively or auto-regressively replace entities in order to ensure semantic consistency w.r.t. the source text.
Experiments show that our models significantly boost the factual consistency of system-generated summaries without sacrificing summary quality in terms of both automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-10-06T02:51:02Z) - Generating (Factual?) Narrative Summaries of RCTs: Experiments with
Neural Multi-Document Summarization [22.611879349101596]
We evaluate modern neural models for abstractive summarization of relevant article abstracts from systematic reviews.
We find that modern summarization systems yield consistently fluent and relevant synopses, but that they are not always factual.
arXiv Detail & Related papers (2020-08-25T22:22:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.