DLGNet-Task: An End-to-end Neural Network Framework for Modeling
Multi-turn Multi-domain Task-Oriented Dialogue
- URL: http://arxiv.org/abs/2010.01693v2
- Date: Tue, 6 Oct 2020 16:31:06 GMT
- Title: DLGNet-Task: An End-to-end Neural Network Framework for Modeling
Multi-turn Multi-domain Task-Oriented Dialogue
- Authors: Oluwatobi O. Olabiyi, Prarthana Bhattarai, C. Bayan Bruss, Zachary
Kulis
- Abstract summary: We present a new framework, DLGNet-Task, a unified task-oriented dialogue system.
Our framework enjoys the controllable, verifiable, and explainable outputs of modular approaches.
Using DLGNet-Task in conversational AI systems reduces the level of effort required for developing, deploying, and maintaining intelligent assistants.
- Score: 0.0
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Task oriented dialogue (TOD) requires the complex interleaving of a number of
individually controllable components with strong guarantees for explainability
and verifiability. This has made it difficult to adopt the multi-turn
multi-domain dialogue generation capabilities of streamlined end-to-end
open-domain dialogue systems. In this paper, we present a new framework,
DLGNet-Task, a unified task-oriented dialogue system which employs
autoregressive transformer networks such as DLGNet and GPT-2/3 to complete user
tasks in multi-turn multi-domain conversations. Our framework enjoys the
controllable, verifiable, and explainable outputs of modular approaches, and
the low development, deployment and maintenance cost of end-to-end systems.
Treating open-domain system components as additional TOD system modules allows
DLGNet-Task to learn the joint distribution of the inputs and outputs of all
the functional blocks of existing modular approaches such as, natural language
understanding (NLU), state tracking, action policy, as well as natural language
generation (NLG). Rather than training the modules individually, as is common
in real-world systems, we trained them jointly with appropriate module
separations. When evaluated on the MultiWOZ2.1 dataset, DLGNet-Task shows
comparable performance to the existing state-of-the-art approaches.
Furthermore, using DLGNet-Task in conversational AI systems reduces the level
of effort required for developing, deploying, and maintaining intelligent
assistants at scale.
Related papers
- Training Zero-Shot Generalizable End-to-End Task-Oriented Dialog System Without Turn-level Dialog Annotations [2.757798192967912]
This work employs multi-task instruction fine-tuning to create more efficient and scalable task-oriented dialogue systems.
Our approach outperforms both state-of-the-art models trained on annotated data and billion-scale parameter off-the-shelf ChatGPT models.
arXiv Detail & Related papers (2024-07-21T04:52:38Z) - Many Hands Make Light Work: Task-Oriented Dialogue System with Module-Based Mixture-of-Experts [9.129081545049992]
Task-oriented dialogue systems have greatly benefited from pre-trained language models (PLMs)
We propose Soft Mixture-of-Expert Task-Oriented Dialogue system (SMETOD)
SMETOD leverages an ensemble of Mixture-of-Experts (MoEs) to excel at subproblems and generate specialized outputs for task-oriented dialogues.
We extensively evaluate our model on three benchmark functionalities: intent prediction, dialogue state tracking, and dialogue response generation.
arXiv Detail & Related papers (2024-05-16T01:02:09Z) - Large Language Models meet Network Slicing Management and Orchestration [0.3644165342767221]
This paper proposes a future vision for network slicing powered by Large Language Models (LLMs) and multi-agent systems.
We discuss the challenges associated with implementing this framework and potential solutions to mitigate them.
arXiv Detail & Related papers (2024-03-20T16:29:52Z) - When Large Language Model Agents Meet 6G Networks: Perception,
Grounding, and Alignment [100.58938424441027]
We propose a split learning system for AI agents in 6G networks leveraging the collaboration between mobile devices and edge servers.
We introduce a novel model caching algorithm for LLMs within the proposed system to improve model utilization in context.
arXiv Detail & Related papers (2024-01-15T15:20:59Z) - InstructTODS: Large Language Models for End-to-End Task-Oriented
Dialogue Systems [60.53276524369498]
Large language models (LLMs) have been used for diverse tasks in natural language processing (NLP)
We present InstructTODS, a novel framework for zero-shot end-to-end task-oriented dialogue systems.
InstructTODS generates a proxy belief state that seamlessly translates user intentions into dynamic queries.
arXiv Detail & Related papers (2023-10-13T06:36:26Z) - Zero-Shot Generalizable End-to-End Task-Oriented Dialog System using
Context Summarization and Domain Schema [2.7178968279054936]
State-of-the-art approaches in task-oriented dialog systems formulate the problem as a conditional sequence generation task.
This requires labeled training data for each new domain or task.
We introduce a novel Zero-Shot generalizable end-to-end Task-oriented Dialog system, ZS-ToD.
arXiv Detail & Related papers (2023-03-28T18:56:31Z) - Can Current Task-oriented Dialogue Models Automate Real-world Scenarios
in the Wild? [48.79943762731801]
Task-oriented dialogue (TOD) systems are mainly based on the slot-filling-based TOD (SF-TOD) framework.
We argue that the current TOD benchmarks are limited to surrogate real-world scenarios and that the current TOD models are still a long way to cover the scenarios.
In WebTOD, the dialogue system learns how to understand the web/mobile interface that the human agent interacts with, powered by a large-scale language model.
arXiv Detail & Related papers (2022-12-20T18:18:41Z) - MULTI3NLU++: A Multilingual, Multi-Intent, Multi-Domain Dataset for
Natural Language Understanding in Task-Oriented Dialogue [115.32009638844059]
We extend the English only NLU++ dataset to include manual translations into a range of high, medium, and low resource languages.
Because of its multi-intent property, MULTI3NLU++ represents complex and natural user goals.
We use MULTI3NLU++ to benchmark state-of-the-art multilingual models for the Natural Language Understanding tasks of intent detection and slot labelling.
arXiv Detail & Related papers (2022-12-20T17:34:25Z) - Enhancing Task Bot Engagement with Synthesized Open-Domain Dialog [89.35658776144638]
It is essential to build a system that can handle both TOD and ODD and access different knowledge sources.
We propose a framework for automatically generating dialogues that combine knowledge-grounded ODDs and TODs in various settings.
We introduce a unified model PivotBot that is capable of appropriately adopting TOD and ODD modes and accessing different knowledge sources.
arXiv Detail & Related papers (2022-12-20T05:51:47Z) - UniConv: A Unified Conversational Neural Architecture for Multi-domain
Task-oriented Dialogues [101.96097419995556]
"UniConv" is a novel unified neural architecture for end-to-end conversational systems in task-oriented dialogues.
We conduct comprehensive experiments in dialogue state tracking, context-to-text, and end-to-end settings on the MultiWOZ2.1 benchmark.
arXiv Detail & Related papers (2020-04-29T16:28:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.