An Empirical Study of Multitask Learning to Improve Open Domain Dialogue
Systems
- URL: http://arxiv.org/abs/2304.08115v1
- Date: Mon, 17 Apr 2023 09:44:56 GMT
- Title: An Empirical Study of Multitask Learning to Improve Open Domain Dialogue
Systems
- Authors: Mehrdad Farahani, Richard Johansson
- Abstract summary: This paper describes an investigation where four different auxiliary tasks are added to small and medium-sized GPT-2 models.
The results show that the introduction of the new auxiliary tasks leads to small but consistent improvement in evaluations of the investigated models.
- Score: 0.13706331473063876
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Autoregressive models used to generate responses in open-domain dialogue
systems often struggle to take long-term context into account and to maintain
consistency over a dialogue. Previous research in open-domain dialogue
generation has shown that the use of \emph{auxiliary tasks} can introduce
inductive biases that encourage the model to improve these qualities. However,
most previous research has focused on encoder-only or encoder/decoder models,
while the use of auxiliary tasks in \emph{decoder-only} autoregressive models
is under-explored. This paper describes an investigation where four different
auxiliary tasks are added to small and medium-sized GPT-2 models fine-tuned on
the PersonaChat and DailyDialog datasets. The results show that the
introduction of the new auxiliary tasks leads to small but consistent
improvement in evaluations of the investigated models.
Related papers
- Promoting Open-domain Dialogue Generation through Learning Pattern
Information between Contexts and Responses [5.936682548344234]
This paper improves the quality of generated responses by learning the implicit pattern information between contexts and responses in the training samples.
We also design a response-aware mechanism for mining the implicit pattern information between contexts and responses so that the generated replies are more diverse and approximate to human replies.
arXiv Detail & Related papers (2023-09-06T08:11:39Z) - Manual-Guided Dialogue for Flexible Conversational Agents [84.46598430403886]
How to build and use dialogue data efficiently, and how to deploy models in different domains at scale can be critical issues in building a task-oriented dialogue system.
We propose a novel manual-guided dialogue scheme, where the agent learns the tasks from both dialogue and manuals.
Our proposed scheme reduces the dependence of dialogue models on fine-grained domain ontology, and makes them more flexible to adapt to various domains.
arXiv Detail & Related papers (2022-08-16T08:21:12Z) - GODEL: Large-Scale Pre-Training for Goal-Directed Dialog [119.1397031992088]
We introduce GODEL, a large pre-trained language model for dialog.
We show that GODEL outperforms state-of-the-art pre-trained dialog models in few-shot fine-tuning setups.
A novel feature of our evaluation methodology is the introduction of a notion of utility that assesses the usefulness of responses.
arXiv Detail & Related papers (2022-06-22T18:19:32Z) - DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for
Dialog Response Generation [80.45816053153722]
DialogVED introduces continuous latent variables into the enhanced encoder-decoder pre-training framework to increase the relevance and diversity of responses.
We conduct experiments on PersonaChat, DailyDialog, and DSTC7-AVSD benchmarks for response generation.
arXiv Detail & Related papers (2022-04-27T16:18:15Z) - Dialogue Summarization with Supporting Utterance Flow Modeling and Fact
Regularization [58.965859508695225]
We propose an end-to-end neural model for dialogue summarization with two novel modules.
The supporting utterance flow modeling helps to generate a coherent summary by smoothly shifting the focus from the former utterances to the later ones.
The fact regularization encourages the generated summary to be factually consistent with the ground-truth summary during model training.
arXiv Detail & Related papers (2021-08-03T03:09:25Z) - RADDLE: An Evaluation Benchmark and Analysis Platform for Robust
Task-oriented Dialog Systems [75.87418236410296]
We introduce the RADDLE benchmark, a collection of corpora and tools for evaluating the performance of models across a diverse set of domains.
RADDLE is designed to favor and encourage models with a strong generalization ability.
We evaluate recent state-of-the-art systems based on pre-training and fine-tuning, and find that grounded pre-training on heterogeneous dialog corpora performs better than training a separate model per domain.
arXiv Detail & Related papers (2020-12-29T08:58:49Z) - Learning an Unreferenced Metric for Online Dialogue Evaluation [53.38078951628143]
We propose an unreferenced automated evaluation metric that uses large pre-trained language models to extract latent representations of utterances.
We show that our model achieves higher correlation with human annotations in an online setting, while not requiring true responses for comparison during inference.
arXiv Detail & Related papers (2020-05-01T20:01:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.