SPACE-2: Tree-Structured Semi-Supervised Contrastive Pre-training for
Task-Oriented Dialog Understanding
- URL: http://arxiv.org/abs/2209.06638v1
- Date: Wed, 14 Sep 2022 13:42:50 GMT
- Title: SPACE-2: Tree-Structured Semi-Supervised Contrastive Pre-training for
Task-Oriented Dialog Understanding
- Authors: Wanwei He, Yinpei Dai, Binyuan Hui, Min Yang, Zheng Cao, Jianbo Dong,
Fei Huang, Luo Si, Yongbin Li
- Abstract summary: We propose a tree-structured pre-trained conversation model, which learns dialog representations from limited labeled dialogs and large-scale unlabeled dialog corpora.
Our method can achieve new state-of-the-art results on the DialoGLUE benchmark consisting of seven datasets and four popular dialog understanding tasks.
- Score: 68.94808536012371
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Pre-training methods with contrastive learning objectives have shown
remarkable success in dialog understanding tasks. However, current contrastive
learning solely considers the self-augmented dialog samples as positive samples
and treats all other dialog samples as negative ones, which enforces dissimilar
representations even for dialogs that are semantically related. In this paper,
we propose SPACE-2, a tree-structured pre-trained conversation model, which
learns dialog representations from limited labeled dialogs and large-scale
unlabeled dialog corpora via semi-supervised contrastive pre-training.
Concretely, we first define a general semantic tree structure (STS) to unify
the inconsistent annotation schema across different dialog datasets, so that
the rich structural information stored in all labeled data can be exploited.
Then we propose a novel multi-view score function to increase the relevance of
all possible dialogs that share similar STSs and only push away other
completely different dialogs during supervised contrastive pre-training. To
fully exploit unlabeled dialogs, a basic self-supervised contrastive loss is
also added to refine the learned representations. Experiments show that our
method can achieve new state-of-the-art results on the DialoGLUE benchmark
consisting of seven datasets and four popular dialog understanding tasks. For
reproducibility, we release the code and data at
https://github.com/AlibabaResearch/DAMO-ConvAI/tree/main/space-2.
Related papers
- KPT: Keyword-guided Pre-training for Grounded Dialog Generation [82.68787152707455]
We propose KPT (guided Pre-Training), a novel self-supervised pre-training method for grounded dialog generation.
Specifically, we use a pre-trained language model to extract the most uncertain tokens in the dialog as keywords.
We conduct extensive experiments on various few-shot knowledge-grounded generation tasks, including grounding on dialog acts, knowledge graphs, persona descriptions, and Wikipedia passages.
arXiv Detail & Related papers (2022-12-04T04:05:01Z) - CGoDial: A Large-Scale Benchmark for Chinese Goal-oriented Dialog
Evaluation [75.60156479374416]
CGoDial is a new challenging and comprehensive Chinese benchmark for Goal-oriented Dialog evaluation.
It contains 96,763 dialog sessions and 574,949 dialog turns totally, covering three datasets with different knowledge sources.
To bridge the gap between academic benchmarks and spoken dialog scenarios, we either collect data from real conversations or add spoken features to existing datasets via crowd-sourcing.
arXiv Detail & Related papers (2022-11-21T16:21:41Z) - SPACE-3: Unified Dialog Model Pre-training for Task-Oriented Dialog
Understanding and Generation [123.37377363355363]
SPACE-3 is a novel unified semi-supervised pre-trained conversation model learning from large-scale dialog corpora.
It can be effectively fine-tuned on a wide range of downstream dialog tasks.
Results show that SPACE-3 achieves state-of-the-art performance on eight downstream dialog benchmarks.
arXiv Detail & Related papers (2022-09-14T14:17:57Z) - OPAL: Ontology-Aware Pretrained Language Model for End-to-End
Task-Oriented Dialogue [40.62090743056549]
This paper presents an ontology-aware pretrained language model (OPAL) for end-to-end task-oriented dialogue (TOD)
Unlike chit-chat dialogue models, task-oriented dialogue models fulfill at least two task-specific modules: dialogue state tracker (DST) and response generator (RG)
arXiv Detail & Related papers (2022-09-10T04:38:27Z) - Structure Extraction in Task-Oriented Dialogues with Slot Clustering [94.27806592467537]
In task-oriented dialogues, dialogue structure has often been considered as transition graphs among dialogue states.
We propose a simple yet effective approach for structure extraction in task-oriented dialogues.
arXiv Detail & Related papers (2022-02-28T20:18:12Z) - GALAXY: A Generative Pre-trained Model for Task-Oriented Dialog with
Semi-Supervised Learning and Explicit Policy Injection [36.77204909711832]
We propose a novel pre-trained dialog model that explicitly learns dialog policy from limited labeled dialogs and large-scale unlabeled dialog corpora.
Specifically, we introduce a dialog act prediction task for policy optimization during pre-training and employ a consistency regularization term to refine the learned representation.
Empirical results show that GALAXY substantially improves the performance of task-oriented dialog systems.
arXiv Detail & Related papers (2021-11-29T15:24:36Z) - Reasoning in Dialog: Improving Response Generation by Context Reading
Comprehension [49.92173751203827]
In multi-turn dialog, utterances do not always take the full form of sentences.
We propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question.
arXiv Detail & Related papers (2020-12-14T10:58:01Z) - Hierarchical Pre-training for Sequence Labelling in Spoken Dialog [10.216901061363641]
We propose a new approach to learn generic representations adapted to spoken dialog.
We obtain our representations with a hierarchical encoder based on transformer architectures.
Pre-training is performed on OpenSubtitles: a large corpus of spoken dialog containing over $2.3$ billion of tokens.
arXiv Detail & Related papers (2020-09-23T13:54:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.