Mind the Gap! Injecting Commonsense Knowledge for Abstractive Dialogue
Summarization
- URL: http://arxiv.org/abs/2209.00930v1
- Date: Fri, 2 Sep 2022 10:08:28 GMT
- Title: Mind the Gap! Injecting Commonsense Knowledge for Abstractive Dialogue
Summarization
- Authors: Seungone Kim, Se June Joo, Hyungjoo Chae, Chaehyeong Kim, Seung-won
Hwang, Jinyoung Yeo
- Abstract summary: We present SICK, a framework that uses commonsense inferences as additional context.
With injected commonsense knowledge, our framework generates more informative and consistent summaries than existing methods.
- Score: 13.863545975204019
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we propose to leverage the unique characteristics of dialogues
sharing commonsense knowledge across participants, to resolve the difficulties
in summarizing them. We present SICK, a framework that uses commonsense
inferences as additional context. Compared to previous work that solely relies
on the input dialogue, SICK uses an external knowledge model to generate a rich
set of commonsense inferences and selects the most probable one with a
similarity-based selection method. Built upon SICK, SICK++ utilizes commonsense
as supervision, where the task of generating commonsense inferences is added
upon summarizing the dialogue in a multi-task learning setting. Experimental
results show that with injected commonsense knowledge, our framework generates
more informative and consistent summaries than existing methods.
Related papers
- DiffuCOMET: Contextual Commonsense Knowledge Diffusion [29.23102821128395]
In this work, we develop a series of knowledge models, DiffuCOMET, that leverage diffusion to learn to reconstruct the implicit semantic connections between narrative contexts and relevant commonsense knowledge.
To evaluate DiffuCOMET, we introduce new metrics for commonsense inference that more closely measure knowledge diversity and contextual relevance.
Our results on two different benchmarks, ComFact and WebNLG+, show that knowledge generated by DiffuCOMET achieves a better trade-off between commonsense diversity, contextual relevance and alignment to known gold references.
arXiv Detail & Related papers (2024-02-26T20:35:34Z) - Generative Knowledge Selection for Knowledge-Grounded Dialogues [40.47433331803577]
We propose a simple yet effective generative approach for knowledge selection, called GenKS.
GenKS learns to select snippets by generating their identifiers with a sequence-to-sequence model.
We conduct experiments on three benchmark datasets, and verify GenKS achieves the best results on both knowledge selection and response generation.
arXiv Detail & Related papers (2023-04-10T19:49:55Z) - Position Matters! Empirical Study of Order Effect in Knowledge-grounded
Dialogue [54.98184262897166]
We investigate how the order of the knowledge set can influence autoregressive dialogue systems' responses.
We propose a simple and novel technique to alleviate the order effect by modifying the position embeddings of knowledge input.
arXiv Detail & Related papers (2023-02-12T10:13:00Z) - elBERto: Self-supervised Commonsense Learning for Question Answering [131.51059870970616]
We propose a Self-supervised Bidirectional Representation Learning of Commonsense framework, which is compatible with off-the-shelf QA model architectures.
The framework comprises five self-supervised tasks to force the model to fully exploit the additional training signals from contexts containing rich commonsense.
elBERto achieves substantial improvements on out-of-paragraph and no-effect questions where simple lexical similarity comparison does not help.
arXiv Detail & Related papers (2022-03-17T16:23:45Z) - Knowledge-Grounded Dialogue Generation with a Unified Knowledge
Representation [78.85622982191522]
Existing systems perform poorly on unseen topics due to limited topics covered in the training data.
We present PLUG, a language model that homogenizes different knowledge sources to a unified knowledge representation.
It can achieve comparable performance with state-of-the-art methods under a fully-supervised setting.
arXiv Detail & Related papers (2021-12-15T07:11:02Z) - Natural Language Processing with Commonsense Knowledge: A Survey [9.634283896785611]
This paper explores the integration of commonsense knowledge into various NLP tasks.
We highlight key methodologies for incorporating commonsense knowledge and their applications across different NLP tasks.
The paper also examines the challenges and emerging trends in enhancing NLP systems with commonsense reasoning.
arXiv Detail & Related papers (2021-08-10T13:25:29Z) - Retrieval-Free Knowledge-Grounded Dialogue Response Generation with
Adapters [52.725200145600624]
We propose KnowExpert to bypass the retrieval process by injecting prior knowledge into the pre-trained language models with lightweight adapters.
Experimental results show that KnowExpert performs comparably with the retrieval-based baselines.
arXiv Detail & Related papers (2021-05-13T12:33:23Z) - Multi-turn Dialogue Reading Comprehension with Pivot Turns and Knowledge [43.352833140317486]
Multi-turn dialogue reading comprehension aims to teach machines to read dialogue contexts and solve tasks such as response selection and answering questions.
This work makes the first attempt to tackle the above two challenges by extracting substantially important turns as pivot utterances.
We propose a pivot-oriented deep selection model (PoDS) on top of the Transformer-based language models for dialogue comprehension.
arXiv Detail & Related papers (2021-02-10T15:00:12Z) - Improving Machine Reading Comprehension with Contextualized Commonsense
Knowledge [62.46091695615262]
We aim to extract commonsense knowledge to improve machine reading comprehension.
We propose to represent relations implicitly by situating structured knowledge in a context.
We employ a teacher-student paradigm to inject multiple types of contextualized knowledge into a student machine reader.
arXiv Detail & Related papers (2020-09-12T17:20:01Z) - Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue [51.513276162736844]
We propose a sequential latent variable model as the first approach to this matter.
The model named sequential knowledge transformer (SKT) can keep track of the prior and posterior distribution over knowledge.
arXiv Detail & Related papers (2020-02-18T11:59:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.