Context-dependent Instruction Tuning for Dialogue Response Generation
- URL: http://arxiv.org/abs/2311.07006v1
- Date: Mon, 13 Nov 2023 01:25:30 GMT
- Title: Context-dependent Instruction Tuning for Dialogue Response Generation
- Authors: Jin Myung Kwak, Minseon Kim, Sung Ju Hwang
- Abstract summary: Recent language models have achieved impressive performance in natural language computation tasks by incorporating instructions with task input during fine-tuning.
We introduce a context-based instruction fine-tuning framework for each multi-turn dialogue.
During the evaluation, the model generates instructions based on the previous context to self-guide the response.
- Score: 61.21790201307179
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent language models have achieved impressive performance in natural
language tasks by incorporating instructions with task input during
fine-tuning. Since all samples in the same natural language task can be
explained with the same task instructions, many instruction datasets only
provide a few instructions for the entire task, without considering the input
of each example in the task. However, this approach becomes ineffective in
complex multi-turn dialogue generation tasks, where the input varies highly
with each turn as the dialogue context changes, so that simple task
instructions cannot improve the generation performance. To address this
limitation, we introduce a context-based instruction fine-tuning framework for
each multi-turn dialogue which generates both responses and instructions based
on the previous context as input. During the evaluation, the model generates
instructions based on the previous context to self-guide the response. The
proposed framework produces comparable or even outstanding results compared to
the baselines by aligning instructions to the input during fine-tuning with the
instructions in quantitative evaluations on dialogue benchmark datasets with
reduced computation budget.
Related papers
- Fine-tuning Large Language Models with Sequential Instructions [2.546845645875049]
We find that existing instruction-tuned models struggle to respond to queries with multiple instructions.
We contend that part of the fine-tuning data mixture should be sequential--containing a chain of interrelated tasks.
We automate this process by turning instructions in existing datasets into diverse and complex sequential instructions.
Models that underwent our sequential instruction tuning show improved results in coding, maths, and open-ended generation.
arXiv Detail & Related papers (2024-03-12T16:33:30Z) - Grounded Complex Task Segmentation for Conversational Assistants [6.188306785668896]
We tackle the recipes domain and convert reading structured instructions into conversational structured ones.
We annotated the structure of instructions according to a conversational scenario, which provided insights into what is expected in this setting.
A further user study showed that users tend to favor steps of manageable complexity and length, and that the proposed methodology can improve the original web-based instructional text.
arXiv Detail & Related papers (2023-09-20T12:55:46Z) - DiactTOD: Learning Generalizable Latent Dialogue Acts for Controllable
Task-Oriented Dialogue Systems [15.087619144902776]
We present a novel end-to-end latent dialogue act model (DiactTOD) that represents dialogue acts in a latent space.
When pre-trained on a large corpus, DiactTOD is able to predict and control dialogue acts to generate controllable responses.
arXiv Detail & Related papers (2023-08-01T23:29:16Z) - Revisiting Conversation Discourse for Dialogue Disentanglement [88.3386821205896]
We propose enhancing dialogue disentanglement by taking full advantage of the dialogue discourse characteristics.
We develop a structure-aware framework to integrate the rich structural features for better modeling the conversational semantic context.
Our work has great potential to facilitate broader multi-party multi-thread dialogue applications.
arXiv Detail & Related papers (2023-06-06T19:17:47Z) - Improving Zero and Few-shot Generalization in Dialogue through
Instruction Tuning [27.92734269206744]
InstructDial is an instruction tuning framework for dialogue.
It consists of a repository of 48 diverse dialogue tasks in a unified text-to-text format created from 59 openly available dialogue datasets.
Our analysis reveals that InstructDial enables good zero-shot performance on unseen datasets and tasks such as dialogue evaluation and intent detection, and even better performance in a few-shot setting.
arXiv Detail & Related papers (2022-05-25T11:37:06Z) - Utterance Rewriting with Contrastive Learning in Multi-turn Dialogue [22.103162555263143]
We introduce contrastive learning and multi-task learning to jointly model the problem.
Our proposed model achieves state-of-the-art performance on several public datasets.
arXiv Detail & Related papers (2022-03-22T10:13:27Z) - Rethinking End-to-End Evaluation of Decomposable Tasks: A Case Study on
Spoken Language Understanding [101.24748444126982]
Decomposable tasks are complex and comprise of a hierarchy of sub-tasks.
Existing benchmarks, however, typically hold out examples for only the surface-level sub-task.
We propose a framework to construct robust test sets using coordinate ascent over sub-task specific utility functions.
arXiv Detail & Related papers (2021-06-29T02:53:59Z) - Reasoning in Dialog: Improving Response Generation by Context Reading
Comprehension [49.92173751203827]
In multi-turn dialog, utterances do not always take the full form of sentences.
We propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question.
arXiv Detail & Related papers (2020-12-14T10:58:01Z) - Learning an Effective Context-Response Matching Model with
Self-Supervised Tasks for Retrieval-based Dialogues [88.73739515457116]
We introduce four self-supervised tasks including next session prediction, utterance restoration, incoherence detection and consistency discrimination.
We jointly train the PLM-based response selection model with these auxiliary tasks in a multi-task manner.
Experiment results indicate that the proposed auxiliary self-supervised tasks bring significant improvement for multi-turn response selection.
arXiv Detail & Related papers (2020-09-14T08:44:46Z) - Masking Orchestration: Multi-task Pretraining for Multi-role Dialogue
Representation Learning [50.5572111079898]
Multi-role dialogue understanding comprises a wide range of diverse tasks such as question answering, act classification, dialogue summarization etc.
While dialogue corpora are abundantly available, labeled data, for specific learning tasks, can be highly scarce and expensive.
In this work, we investigate dialogue context representation learning with various types unsupervised pretraining tasks.
arXiv Detail & Related papers (2020-02-27T04:36:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.