Leveraging Non-dialogue Summaries for Dialogue Summarization
- URL: http://arxiv.org/abs/2210.09474v1
- Date: Mon, 17 Oct 2022 23:34:31 GMT
- Title: Leveraging Non-dialogue Summaries for Dialogue Summarization
- Authors: Seongmin Park, Dongchan Shin, Jihwa Lee
- Abstract summary: We apply transformations to document summarization data pairs to create training data that better befit dialogue summarization.
We conduct extensive experiments across both English and Korean to verify our approach.
- Score: 1.0742675209112622
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To mitigate the lack of diverse dialogue summarization datasets in academia,
we present methods to utilize non-dialogue summarization data for enhancing
dialogue summarization systems. We apply transformations to document
summarization data pairs to create training data that better befit dialogue
summarization. The suggested transformations also retain desirable properties
of non-dialogue datasets, such as improved faithfulness to the source text. We
conduct extensive experiments across both English and Korean to verify our
approach. Although absolute gains in ROUGE naturally plateau as more dialogue
summarization samples are introduced, utilizing non-dialogue data for training
significantly improves summarization performance in zero- and few-shot settings
and enhances faithfulness across all training regimes.
Related papers
- SPECTRUM: Speaker-Enhanced Pre-Training for Long Dialogue Summarization [48.284512017469524]
Multi-turn dialogues are characterized by their extended length and the presence of turn-taking conversations.
Traditional language models often overlook the distinct features of these dialogues by treating them as regular text.
We propose a speaker-enhanced pre-training method for long dialogue summarization.
arXiv Detail & Related papers (2024-01-31T04:50:00Z) - Multijugate Dual Learning for Low-Resource Task-Oriented Dialogue System [40.072084239764465]
We discover that data utilization efficiency in low-resource scenarios can be enhanced by mining alignment information uncertain utterance and deterministic dialogue state.
We innovatively implement dual learning in task-oriented dialogues to exploit the correlation of heterogeneous data.
arXiv Detail & Related papers (2023-05-25T14:38:05Z) - Pre-training Multi-party Dialogue Models with Latent Discourse Inference [85.9683181507206]
We pre-train a model that understands the discourse structure of multi-party dialogues, namely, to whom each utterance is replying.
To fully utilize the unlabeled data, we propose to treat the discourse structures as latent variables, then jointly infer them and pre-train the discourse-aware model.
arXiv Detail & Related papers (2023-05-24T14:06:27Z) - Few-Shot Dialogue Summarization via Skeleton-Assisted Prompt Transfer in
Prompt Tuning [47.336815771549524]
Skeleton-Assisted Prompt Transfer improves prompt transfer from dialogue state tracking to dialogue summarization.
We propose a novel approach with perturbation-based probes requiring neither annotation effort nor domain knowledge.
In-depth analyses demonstrate the effectiveness of our method in facilitating cross-task knowledge transfer in few-shot dialogue summarization.
arXiv Detail & Related papers (2023-05-20T03:32:48Z) - DIONYSUS: A Pre-trained Model for Low-Resource Dialogue Summarization [127.714919036388]
DIONYSUS is a pre-trained encoder-decoder model for summarizing dialogues in any new domain.
Our experiments show that DIONYSUS outperforms existing methods on six datasets.
arXiv Detail & Related papers (2022-12-20T06:21:21Z) - Analyzing and Evaluating Faithfulness in Dialogue Summarization [67.07947198421421]
We first perform the fine-grained human analysis on the faithfulness of dialogue summaries and observe that over 35% of generated summaries are faithfully inconsistent respective the source dialogues.
We present a new model-level faithfulness evaluation method. It examines generation models with multi-choice questions created by rule-based transformations.
arXiv Detail & Related papers (2022-10-21T07:22:43Z) - Enhancing Semantic Understanding with Self-supervised Methods for
Abstractive Dialogue Summarization [4.226093500082746]
We introduce self-supervised methods to compensate shortcomings to train a dialogue summarization model.
Our principle is to detect incoherent information flows using pretext dialogue text to enhance BERT's ability to contextualize the dialogue text representations.
arXiv Detail & Related papers (2022-09-01T07:51:46Z) - Dialogue Summarization with Supporting Utterance Flow Modeling and Fact
Regularization [58.965859508695225]
We propose an end-to-end neural model for dialogue summarization with two novel modules.
The supporting utterance flow modeling helps to generate a coherent summary by smoothly shifting the focus from the former utterances to the later ones.
The fact regularization encourages the generated summary to be factually consistent with the ground-truth summary during model training.
arXiv Detail & Related papers (2021-08-03T03:09:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.