On Task-Adaptive Pretraining for Dialogue Response Selection
- URL: http://arxiv.org/abs/2210.04073v1
- Date: Sat, 8 Oct 2022 17:58:49 GMT
- Title: On Task-Adaptive Pretraining for Dialogue Response Selection
- Authors: Tzu-Hsiang Lin, Ta-Chung Chi, Anna Rumshisky
- Abstract summary: This paper aims to verify assumptions made in previous advancements and understand the source of improvements for dialogue response selection (DRS)
We show that initializing with RoBERTa achieve similar performance as BERT, and anticipate+NSP can outperform all previously proposed TAP tasks.
Additional analyses shows that the main source of improvements comes from the TAP step, and that the NSP task is crucial to DRS.
- Score: 9.502775168613589
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent advancements in dialogue response selection (DRS) are based on the
\textit{task-adaptive pre-training (TAP)} approach, by first initializing their
model with BERT~\cite{devlin-etal-2019-bert}, and adapt to dialogue data with
dialogue-specific or fine-grained pre-training tasks. However, it is uncertain
whether BERT is the best initialization choice, or whether the proposed
dialogue-specific fine-grained learning tasks are actually better than MLM+NSP.
This paper aims to verify assumptions made in previous works and understand the
source of improvements for DRS. We show that initializing with RoBERTa achieve
similar performance as BERT, and MLM+NSP can outperform all previously proposed
TAP tasks, during which we also contribute a new state-of-the-art on the Ubuntu
corpus. Additional analyses shows that the main source of improvements comes
from the TAP step, and that the NSP task is crucial to DRS, different from
common NLU tasks.
Related papers
- InstructTODS: Large Language Models for End-to-End Task-Oriented
Dialogue Systems [60.53276524369498]
Large language models (LLMs) have been used for diverse tasks in natural language processing (NLP)
We present InstructTODS, a novel framework for zero-shot end-to-end task-oriented dialogue systems.
InstructTODS generates a proxy belief state that seamlessly translates user intentions into dynamic queries.
arXiv Detail & Related papers (2023-10-13T06:36:26Z) - Task-Optimized Adapters for an End-to-End Task-Oriented Dialogue System [0.0]
We propose an End-to-end TOD system with Task-d Adapters which learn independently per task, adding only small number of parameters after fixed layers of pre-trained network.
Our method is a model-agnostic approach and does not require prompt-tuning as only input data without a prompt.
arXiv Detail & Related papers (2023-05-04T00:17:49Z) - Towards Unified Prompt Tuning for Few-shot Text Classification [47.71344780587704]
We present the Unified Prompt Tuning (UPT) framework, leading to better few-shot text classification for BERT-style models.
In UPT, a novel paradigm Prompt-Options-Verbalizer is proposed for joint prompt learning across different NLP tasks.
We also design a self-supervised task named Knowledge-enhanced Selective Masked Language Modeling to improve the PLM's generalization abilities.
arXiv Detail & Related papers (2022-05-11T07:40:45Z) - In-Context Learning for Few-Shot Dialogue State Tracking [55.91832381893181]
We propose an in-context (IC) learning framework for few-shot dialogue state tracking (DST)
A large pre-trained language model (LM) takes a test instance and a few annotated examples as input, and directly decodes the dialogue states without any parameter updates.
This makes the LM more flexible and scalable compared to prior few-shot DST work when adapting to new domains and scenarios.
arXiv Detail & Related papers (2022-03-16T11:58:24Z) - Multi-Task Pre-Training for Plug-and-Play Task-Oriented Dialogue System [26.837972034630003]
PPTOD is a unified plug-and-play model for task-oriented dialogue.
We extensively test our model on three benchmark TOD tasks, including end-to-end dialogue modelling, dialogue state tracking, and intent classification.
arXiv Detail & Related papers (2021-09-29T22:02:18Z) - CINS: Comprehensive Instruction for Few-shot Learning in Task-oriented
Dialog Systems [56.302581679816775]
This paper proposes Comprehensive Instruction (CINS) that exploits PLMs with task-specific instructions.
We design a schema (definition, constraint, prompt) of instructions and their customized realizations for three important downstream tasks in ToD.
Experiments are conducted on these ToD tasks in realistic few-shot learning scenarios with small validation data.
arXiv Detail & Related papers (2021-09-10T03:23:06Z) - Task-specific Objectives of Pre-trained Language Models for Dialogue
Adaptation [79.0866650271659]
Common process of utilizing PrLMs is first pre-training on large-scale general corpora with task-independent LM training objectives, then fine-tuning on task datasets with task-specific training objectives.
We introduce task-specific pre-training on in-domain task-related corpora with task-specific objectives.
This procedure is placed between the original two stages to enhance the model understanding capacity of specific tasks.
arXiv Detail & Related papers (2020-09-10T16:46:46Z) - A Tailored Pre-Training Model for Task-Oriented Dialog Generation [60.05269529832447]
We propose a Pre-trained Role Alternating Language model (PRAL) for task-oriented conversational systems.
We introduce a task-oriented dialog pretraining dataset by cleaning 13 existing data sets.
The results show that PRAL performs better or on par with state-of-the-art methods.
arXiv Detail & Related papers (2020-04-24T09:25:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.