A Simple and Efficient Multi-Task Learning Approach for Conditioned
Dialogue Generation
- URL: http://arxiv.org/abs/2010.11140v2
- Date: Sat, 24 Apr 2021 14:51:24 GMT
- Title: A Simple and Efficient Multi-Task Learning Approach for Conditioned
Dialogue Generation
- Authors: Yan Zeng and Jian-Yun Nie
- Abstract summary: Conditioned dialogue generation suffers from the scarcity of labeled responses.
We propose a multi-task learning approach to leverage both labeled dialogue and text data.
- Score: 23.828348485513043
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Conditioned dialogue generation suffers from the scarcity of labeled
responses. In this work, we exploit labeled non-dialogue text data related to
the condition, which are much easier to collect. We propose a multi-task
learning approach to leverage both labeled dialogue and text data. The 3 tasks
jointly optimize the same pre-trained Transformer -- conditioned dialogue
generation task on the labeled dialogue data, conditioned language encoding
task and conditioned language generation task on the labeled text data.
Experimental results show that our approach outperforms the state-of-the-art
models by leveraging the labeled texts, and it also obtains larger improvement
in performance comparing to the previous methods to leverage text data.
Related papers
- DFlow: Diverse Dialogue Flow Simulation with Large Language Models [16.209331014315463]
This paper proposes a novel data augmentation method designed to enhance the diversity of synthetic dialogues.
We generate a task-oriented dialogue dataset comprising 3,886 dialogue flows across 15 different domains.
arXiv Detail & Related papers (2024-10-18T20:35:28Z) - Weakly Supervised Data Augmentation Through Prompting for Dialogue
Understanding [103.94325597273316]
We present a novel approach that iterates on augmentation quality by applying weakly-supervised filters.
We evaluate our methods on the emotion and act classification tasks in DailyDialog and the intent classification task in Facebook Multilingual Task-Oriented Dialogue.
For DailyDialog specifically, using 10% of the ground truth data we outperform the current state-of-the-art model which uses 100% of the data.
arXiv Detail & Related papers (2022-10-25T17:01:30Z) - OPAL: Ontology-Aware Pretrained Language Model for End-to-End
Task-Oriented Dialogue [40.62090743056549]
This paper presents an ontology-aware pretrained language model (OPAL) for end-to-end task-oriented dialogue (TOD)
Unlike chit-chat dialogue models, task-oriented dialogue models fulfill at least two task-specific modules: dialogue state tracker (DST) and response generator (RG)
arXiv Detail & Related papers (2022-09-10T04:38:27Z) - Curriculum-Based Self-Training Makes Better Few-Shot Learners for
Data-to-Text Generation [56.98033565736974]
We propose Curriculum-Based Self-Training (CBST) to leverage unlabeled data in a rearranged order determined by the difficulty of text generation.
Our method can outperform fine-tuning and task-adaptive pre-training methods, and achieve state-of-the-art performance in the few-shot setting of data-to-text generation.
arXiv Detail & Related papers (2022-06-06T16:11:58Z) - Reasoning in Dialog: Improving Response Generation by Context Reading
Comprehension [49.92173751203827]
In multi-turn dialog, utterances do not always take the full form of sentences.
We propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question.
arXiv Detail & Related papers (2020-12-14T10:58:01Z) - Conditioned Text Generation with Transfer for Closed-Domain Dialogue
Systems [65.48663492703557]
We show how to optimally train and control the generation of intent-specific sentences using a conditional variational autoencoder.
We introduce a new protocol called query transfer that allows to leverage a large unlabelled dataset.
arXiv Detail & Related papers (2020-11-03T14:06:10Z) - Modeling Long Context for Task-Oriented Dialogue State Generation [51.044300192906995]
We propose a multi-task learning model with a simple yet effective utterance tagging technique and a bidirectional language model.
Our approaches attempt to solve the problem that the performance of the baseline significantly drops when the input dialogue context sequence is long.
In our experiments, our proposed model achieves a 7.03% relative improvement over the baseline, establishing a new state-of-the-art joint goal accuracy of 52.04% on the MultiWOZ 2.0 dataset.
arXiv Detail & Related papers (2020-04-29T11:02:25Z) - Learning to Select Bi-Aspect Information for Document-Scale Text Content
Manipulation [50.01708049531156]
We focus on a new practical task, document-scale text content manipulation, which is the opposite of text style transfer.
In detail, the input is a set of structured records and a reference text for describing another recordset.
The output is a summary that accurately describes the partial content in the source recordset with the same writing style of the reference.
arXiv Detail & Related papers (2020-02-24T12:52:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.