DialogUSR: Complex Dialogue Utterance Splitting and Reformulation for
Multiple Intent Detection
- URL: http://arxiv.org/abs/2210.11279v1
- Date: Thu, 20 Oct 2022 13:56:35 GMT
- Title: DialogUSR: Complex Dialogue Utterance Splitting and Reformulation for
Multiple Intent Detection
- Authors: Haoran Meng, Zheng Xin, Tianyu Liu, Zizhen Wang, He Feng, Binghuai
Lin, Xuemin Zhao, Yunbo Cao and Zhifang Sui
- Abstract summary: Instead of training a dedicated multi-intent detection model, we propose DialogUSR.
DialogUSR splits multi-intent user query into several single-intent sub-queries.
It then recovers all the coreferred and omitted information in the sub-queries.
- Score: 27.787807111516706
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: While interacting with chatbots, users may elicit multiple intents in a
single dialogue utterance. Instead of training a dedicated multi-intent
detection model, we propose DialogUSR, a dialogue utterance splitting and
reformulation task that first splits multi-intent user query into several
single-intent sub-queries and then recovers all the coreferred and omitted
information in the sub-queries. DialogUSR can serve as a plug-in and
domain-agnostic module that empowers the multi-intent detection for the
deployed chatbots with minimal efforts. We collect a high-quality naturally
occurring dataset that covers 23 domains with a multi-step crowd-souring
procedure. To benchmark the proposed dataset, we propose multiple action-based
generative models that involve end-to-end and two-stage training, and conduct
in-depth analyses on the pros and cons of the proposed baselines.
Related papers
- DialCLIP: Empowering CLIP as Multi-Modal Dialog Retriever [83.33209603041013]
We propose a parameter-efficient prompt-tuning method named DialCLIP for multi-modal dialog retrieval.
Our approach introduces a multi-modal context generator to learn context features which are distilled into prompts within the pre-trained vision-language model CLIP.
To facilitate various types of retrieval, we also design multiple experts to learn mappings from CLIP outputs to multi-modal representation space.
arXiv Detail & Related papers (2024-01-02T07:40:12Z) - Multi-User MultiWOZ: Task-Oriented Dialogues among Multiple Users [51.34484827552774]
We release the Multi-User MultiWOZ dataset: task-oriented dialogues among two users and one agent.
These dialogues reflect interesting dynamics of collaborative decision-making in task-oriented scenarios.
We propose a novel task of multi-user contextual query rewriting: to rewrite a task-oriented chat between two users as a concise task-oriented query.
arXiv Detail & Related papers (2023-10-31T14:12:07Z) - EM Pre-training for Multi-party Dialogue Response Generation [86.25289241604199]
In multi-party dialogues, the addressee of a response utterance should be specified before it is generated.
We propose an Expectation-Maximization (EM) approach that iteratively performs the expectation steps to generate addressee labels.
arXiv Detail & Related papers (2023-05-21T09:22:41Z) - DIONYSUS: A Pre-trained Model for Low-Resource Dialogue Summarization [127.714919036388]
DIONYSUS is a pre-trained encoder-decoder model for summarizing dialogues in any new domain.
Our experiments show that DIONYSUS outperforms existing methods on six datasets.
arXiv Detail & Related papers (2022-12-20T06:21:21Z) - "Think Before You Speak": Improving Multi-Action Dialog Policy by
Planning Single-Action Dialogs [33.78889030078026]
Multi-action dialog policy (MADP) generates multiple atomic dialog actions per turn.
We propose Planning Enhanced Dialog Policy (PEDP), a novel multi-task learning framework that learns single-action dialog dynamics.
Our fully supervised learning-based method achieves a solid task success rate of 90.6%, improving 3% compared to the state-of-the-art methods.
arXiv Detail & Related papers (2022-04-25T07:55:53Z) - UniConv: A Unified Conversational Neural Architecture for Multi-domain
Task-oriented Dialogues [101.96097419995556]
"UniConv" is a novel unified neural architecture for end-to-end conversational systems in task-oriented dialogues.
We conduct comprehensive experiments in dialogue state tracking, context-to-text, and end-to-end settings on the MultiWOZ2.1 benchmark.
arXiv Detail & Related papers (2020-04-29T16:28:22Z) - Multi-Domain Dialogue Acts and Response Co-Generation [34.27525685962274]
We propose a neural co-generation model that generates dialogue acts and responses concurrently.
Our model achieves very favorable improvement over several state-of-the-art models in both automatic and human evaluations.
arXiv Detail & Related papers (2020-04-26T12:21:17Z) - Dialogue-Based Relation Extraction [53.2896545819799]
We present the first human-annotated dialogue-based relation extraction (RE) dataset DialogRE.
We argue that speaker-related information plays a critical role in the proposed task, based on an analysis of similarities and differences between dialogue-based and traditional RE tasks.
Experimental results demonstrate that a speaker-aware extension on the best-performing model leads to gains in both the standard and conversational evaluation settings.
arXiv Detail & Related papers (2020-04-17T03:51:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.